Tags: safety

7

sparkline

Monday, September 2nd, 2024

Does AI benefit the world? – Chelsea Troy

Our ethical struggle with generative models derives in part from the fact that we…sort of can’t have them ethically, right now, to be honest. We have known how to build models like this for a long time, but we did not have the necessary volume of parseable data available until recently—and even then, to get it, companies have to plunder the internet. Sitting around and waiting for consent from all the parties that wrote on the internet over the past thirty years probably didn’t even cross Sam Altman’s mind.

On the environmental front, fans of generative model technology insist that eventually we’ll possess sufficiently efficient compute power to train and run these models without the massive carbon footprint. That is not the case at the moment, and we don’t have a concrete timeline for it. Again, wait around for a thing we don’t have yet doesn’t appeal to investors or executives.

Tuesday, May 21st, 2024

The Lunacy of Artemis (Idle Words)

Maciej rips NASA’s Artemis programme a new one:

Advocates for Artemis insist that the program is more than Apollo 2.0. But as we’ll see, Artemis can’t even measure up to Apollo 1.0. It costs more, does less, flies less frequently, and exposes crews to risks that the steely-eyed missile men of the Apollo era found unacceptable. It’s as if Ford in 2024 released a new model car that was slower, more accident-prone, and ten times more expensive than the Model T.

When a next-generation lunar program can’t meet the cost, performance, or safety standards set three generations earlier, something has gone seriously awry.

Wednesday, May 15th, 2024

AI Safety for Fleshy Humans: a whirlwind tour

This is a terrificly entertaining level-headed in-depth explanation of AI safety. By the end of this year, all three parts will be published; right now the first part is ready for you to read and enjoy.

This 3-part series is your one-stop-shop to understand the core ideas of AI & AI Safety — explained in a friendly, accessible, and slightly opinionated way!

( Related phrases: AI Risk, AI X-Risk, AI Alignment, AI Ethics, AI Not-Kill-Everyone-ism. There is no consensus on what these phrases do & don’t mean, so I’m just using “AI Safety” as a catch-all.)

Monday, May 13th, 2024

Manifesto for a Humane Web

I endorse this message.

This manifesto is intended as a personal response to the current state of the web. It is a statement of intent and a call to arms, inviting you, the reader, to go forth and build humane websites, and to resist the erosion of the web we know and love.

Thursday, February 23rd, 2023

Privacy in the product design lifecycle | ICO

A very handy guide to considering privacy at all stages of digital product design:

This guidance is written for technology professionals such as product and UX designers, software engineers, QA testers, and product managers.

  1. The case for privacy
  2. Privacy in the kick-off stage
  3. Privacy in the research stage
  4. Privacy in the design stage
  5. Privacy in the development stage
  6. Privacy in the launch phase
  7. Privacy in the post-launch phase

Wednesday, February 8th, 2023

Privacy, Seriously | ICO

This looks like an excellent—and free!—online event centred on privacy and safety. It’s got Eva PenzeyMoog, Robin Berjon and more!

Monday, May 4th, 2020

ongoing by Tim Bray · Bye, Amazon

May 1st was my last day as a VP and Distinguished Engineer at Amazon Web Services, after five years and five months of rewarding fun. I quit in dismay at Amazon firing whistleblowers who were making noise about warehouse employees frightened of Covid-19.

Fair play, Tim Bray!

The victims weren’t abstract entities but real people; here are some of their names: Courtney Bowden, Gerald Bryson, Maren Costa, Emily Cunningham, Bashir Mohammed, and Chris Smalls.

I’m sure it’s a coincidence that every one of them is a person of color, a woman, or both. Right?