Events2Join

AI companies are not on track to secure model weights — EA Forum


AI companies are not on track to secure model weights — EA Forum

I think AI companies are not currently on track to secure model weights. The good news is, I don't think we have to solve any new fundamental problems in ...

Effective Altruism News on X: "AI companies are not on track to ...

AI companies are not on track to secure model weights — Effective Altruism Forum https://t.co/nVOMKCFHkL.

EA Forum Posts on X: "New popular post from the EA Forum: "AI ...

New popular post from the EA Forum: "AI companies are not on track to secure model weights" by Jeffrey Ladish https://t.co/rKBB9c7kzY.

Information security considerations for AI and the long term future

EA Forum · GIVING SEASON 2024. Login Sign up. Hide table of ... AI companies are not on track to secure model weights · Jeffrey Ladish.

Jeffrey Ladish - EA Forum

5. Bounty for Evidence on Some of Palisade Research's Beliefs · 1mo · 1m read. 0 ; 73. AI companies are not on track to secure model weights · 4mo · 23m read. 3 ; 23.

The Precipice Revisited — EA Forum

I'm going to dive into four of the biggest risks — climate change, nuclear, pandemics, and AI — to show how they've changed ...

AI Governance Needs Technical Work - Effective Altruism Forum

It would be bad if people steal unsafe ML models and deploy them. It would also be bad if AI developers rush to deploy their own models (e.g. ...

[DISCUSSION] How much can we trust OpenAI (and other large AI ...

tldr: Do you trust OpenAI or other large AI companies with your data? Do you reckon it's just a matter of time before they find all of the ...

Effective Altruism Global - EA Forum

Do not use this tag for EA Global talks, unless the talks themselves are ... AI companies are not on track to secure model weights · Jeffrey Ladish. + 0 ...

My thoughts on the social response to AI risk - Effective Altruism Forum

AI safety, including the problem of having AIs not kill everyone, is a natural thing for people to care about. Now, I don't know exactly what ...

Artificial Intelligence Index Report 2023 - Stanford University

In 2022, there were 32 significant industry-produced machine learning models compared to just three produced by academia. Building state-of-the-art AI systems.

Jeffrey Ladish's Posts - Effective Altruism forum viewer

A faster way to browse Effective Altruism Forum. ... AI companies are not on track to secure model weights ... EA Hangout Prisoners' Dilemma · Jeffrey Ladish9 ...

Harnessing Artificial Intelligence to Meet Global Challenges

In climate science, AI models are starting to enhance weather prediction, as well as advancing whole-earth models for water management, ...

Ethics and governance of artificial intelligence for health - IRIS Home

Several large technology companies, through the Frontier Model Forum, have committed ... AI weights are not “open source”. Open Core Ventures, 27 June 2023 ...

AI Weights: Securing the Heart and Soft Underbelly of Artificial ...

It is essential that these weights are protected from bad actors. As we move towards greater AI integration in business, model weights become ...

Preventing an AI-related catastrophe - 80,000 Hours

And if the technology keeps advancing at this pace, it seems clear there will be major effects on society. At the very least, automating tasks makes carrying ...

Investigating the Influence of Artificial Intelligence on Business ...

For organizations, the development of new business models and competitive advantages through the integration of artificial intelligence (AI) in business and ...

Counterarguments to the basic AI risk case — EA Forum

We don't know of any procedure to do it, and we have theoretical reasons to expect that AI systems produced through machine learning training ...

EA Forum Podcast (Curated & popular) | Podcast on Podbay

US AGI research must employ vastly better cybersecurity, to protect both model weights and algorithmic secrets. ... “I bet Greg Colbourn 10 k€ that AI will not ...

Why Anthropic and OpenAI are obsessed with securing LLM model ...

Why AI model companies are concerned about malicious actors accessing the weights of the most sophisticated and powerful LLMs.