Discrimination by law enforcement, exclusion from loan approval, bad moderation on social networking, cheating on exams, creating fake news or media about people, swallowing up user data... all the negative social impact of AI can be achieved without it, and much of it is already illegal anyway.
Legislation that is predicated on AI will fail in the long run. Legislation that focuses on the actual negative outcomes will stand the test of time much more.
With the big differences being massive automatisation, huge reduction of cost and no one to blame when things go wrong... It's like saying a nuke and a knife are the same because they both kill
For example things like privacy and surveillance laws obviously need updating in the face of advances in networking, data collection at scale, etc. Same with copyright in the face of plentiful copying.
But good laws will as you say address what is now possible or dangerous, as opposed to any specific implementation or general purpose technology involved. The tech just sets the context for what protections are needed.
Are you working on fixing those root problems? Or after dismissing short term policy bandaids, are you going to go back to working in an industry where you will probably make more money in the short run if governments don't do any tech regulation in the short run?
Your commitment to the long run will lead to paralysis and do nothing in the long run.
But most of the pushback I've seen to AI in policy is so over-fit to current AI that it would be trivial to work around it. You can argue that we'd be letting perfect be the enemy of good, but I think we'd be making policies that will be out of date by the time they even make it into law, and that we'll never make any progress at all.
That said, I'm all for being proven wrong. The US tends to write highly specific legislation so I'm sure it'll try a few of these. The EU tends to write much more vague legislation specifically for this reason. We'll see how they end up working.
Yeah we better let these important topics in the hands of very stable people like Musk or Thiel, they for sure know what the people want
> make more money in the short run if governments don't do any tech regulation in the short run?
"Money money money money", homo sapiens decerebration under capitalism is quite something to witness. Maybe just maybe there is more to life than raw productivity and money... The root causes you're talking about are greed and an unbound quest for "progress", piling more in top will certainly not help
What skills won't be replaced? The only ones I can think of either have a large physical component, or are only doable by a tiny fraction of the current workforce.
As for the ones with a physical component (plumbers being the most cited), the cognitive parts of the job (the "skilled" part of skilled labor) can be replaced while having the person just following directions demonstrated onscreen for them. And of course, the robots aren't far behind, since the main hard part of making a capable robot is the AI part.
Robots are far behind.
Mechanical hands with human equivalent performance is as hard as the AI part.
Strong, fast, durable, tough, touch and temp sensitive, dexterous, light, water-proof, energy efficient, non-overheating.
Muscles and tendons in human hands and forearms self-heal and grow stronger with more use.
Mechanical tendons stretch and break. Small motors have plenty of issues of their own.
As a professional robotics engineer I can tell you for a fact they are coming soon.
But AI can't be held liable for its actions, that is one role. It has no direct access to the context it is working in, so it needs humans as a bridge. In the end AI produce outcomes in the same local context, which is for the user. So from intent to guidance to outcomes they are all user based, costs and risks too.
I find it pessimistic to take that static view on work, as if "that's it, all we needed is invented", and now we are fighting for positions like musical chairs
Daily reminder that the vast majority of value generated by productivity boost brought by technology in the last 50 years doesn't benefit the workers
https://encrypted-tbn0.gstatic.com/images?q=tbn:ANd9GcSG4s-x...
How seriously would you take a proposal on car pollution regulation and traffic law updates written by Volkswagen?
they more or less wrote the EU emission regulations
the only reason diesel cars were sold in huge numbers in the EU
but at least a couple of these proposals seem to boil down to needing to tax the absolute crap out of the AI companies. which seems pretty obviously true, and its interesting that the ai companies are already saying that.
This is just cheap PR to launder legitimacy and urgency. To create false equivalence between AI agent and an employee.
I think this is a sign of weakness, having seen AI rolled out in many companies where it already shows signs of being absolute disaster (like summaries changing meaning and losing important details - so tasks go in wrong direction and take time to be corrected, developers creating unprecedented amount of tech debt with their vibe coded features, massive amount of content that sound important, but it is just equivalent of spam, managers spending ours with LLM "researching" strategy feeding the FOMO and so on).
I can't speak to this particular proposal or the motivations behind it, but I think my approach is the smart play in the present circumstances. Why publish something brazenly self-serving that will at best be forgotten two weeks later, or at worst be added to the list of reasons a bunch of people have to hate you, when you could instead earn some goodwill as a benevolent thought leader and maybe get some academics and politicians to come out of the woodwork backing your ideas?
If the industry is successful and a particular player doesn't fall behind the competition, they're going to be making obscene amounts of money regardless. Better to have a happy and successful public that can't imagine life without you than a public in Great-Depression-like conditions that wants you dead and will only vote for politicians who campaign on banning your product.
As an aside, I'm not sold on the idea of taxes that specifically increase the cost of AI. I don't think it's wise to disincentivize AI usage or artificially inflate costs. (That would particularly hurt anyone with use cases that aren't connected to immediate profit.) If AI has the impact most of us would like it to have, the economy will become way more productive and the public will get its share of that through corporate taxes anyway. I'd rather just close tax loopholes and start laying the groundwork for a future system of distributing resources in a post-employment world.
My current preference is a guaranteed educational/training stipend for any unemployed adult who wants one, and changing the standard career advice for the next generation from "learn to code" to "learn to startup". Looking forward a decade from now, if employment as we know it is scarce, but the economy is flush with capital and automated labor is dirt cheap, it seems to me that self-employment will reemerge as the dominant career path — and anyone who can't raise funding for their business (or acquire grants for their research) will simply need to keep leveling up until they can. Maybe eventually we'll have the resources to transition to a full UBI, but in the meantime, we'd need a transitional system that could provide for the unemployed masses without incentivizing everyone else to suddenly quit jobs that were still necessary. Just my 2c.
I don’t trust them. Their strategy is to say “don’t worry about all your jobs being taken by our technology. We (AI companies) are going to be taxed so much that you are going to be living a wealthy and fruitful life making meme photos and looking at AI porn. Don’t be concerned about how you’ll pay your bills. We’ll work it all out. Trust us.”
As the masses fade into permanent unemployment, this will likely coincide with (and be partially caused by) a corresponding proliferation in intelligent humanoid robots.
At a certain point, "turning on them" becomes physically impossible.
LLM-Attention centric AI isn’t the end of AI development
So if they are successful at locking in it will be at their own demise because it doesn’t cover the infinity many pathways for AI to continue down, specifically intersections with robotics and physical manipulation, that are ultimately way more impactful on society.
Until the plurality of humans on the earth understand that human exceptionalism is no longer something to be taking for granted (and shouldn’t have been) there’s never going to be effective global governance of technology.
Could you elaborate more on this? FYI fully agreed on the former sentences.
The tasks humans are best at now are different than 10kya.
The world changes, new human jobs are made and humans collectively move up the abstraction chain. Schumpeter called this creative destruction and “capital + technology” is the transition function.
At the point where “capital + technology” does not need a human anymore and that will happen (if not in my lifetime then at least in the next 500 years) then there will be nothing more to argue for or retain.
So unless humanity recognizes this and decides to organize as humans (not as europeans, or alabamans or han etc…) then this is the only possible outcome.
Me personally, I don’t think that’s mathematically/energetically possible for humans to do because we’re not biologically capable of that level of eusocial coordination.
Maybe you are alcohol, gambling, and pornography resistant but maybe you have friends and family that aren't. Are you picking up their slack?
What circumstances make "going Amish" look, not just reasonable, but necessary for survival?
High school version: https://kemendo.com/basiccohesion.html
Full draft PDF https://kemendo.com/GTC.pdf
Without any modifications - MOOCs have single digit completion rates. This is high quality, free, publicly available educational material.
The vast majority of people do not simply have the time, money, or undivided attention - to get a new domain under their belt.
This is “help miners learn code” territory.
Not sure MOOCs can be taken as an useful alibi to measure success of upskill. Most (employers) won't honor the MOCC certs, and people do MOOC while working. Taking a MOOC doesn't inherently ensure that the learner has mastered the course they took, hence there is less incentive in completing too.
1. https://www.foreignaffairs.com/united-states/coming-ai-backl...
I’m aware that this level of nihilism is difficult for many to stomach - but it’s only nihilism if you believe in fairy tales.
If we are talking about reality then we have to deal with the impossible challenges we are facing.
The fact is that retraining will not work - nor is this the first time it’s been named as a hope up in the past 25 years. (“teaching miners how to code” comes from the last time this was a big hope in America.)
If it helps you feel better - I said unmodified MOOCs.
With some changes, you can increase MOOC completion rates - but there isn’t enough lift even after that.
I used to be a champion for education initiatives to up skill workers - from a time before MOOCs.
The failure of MOOCs was the end of that hope because it showed there was a gape between the ideal and reality.
People simply can’t retrain like that.
Financial circularity could also lead to instability.
I hope people will eventually revisit these predictions and admit they were wrong.
Incredible stuff...
Proposal written by billionaire trying to shift taxation even more away from themselves and even more to everyone else.
> Accelerate permits and approvals for AI infrastructure
Oh, they want that? Who would not say.
The reasons for that will be proposed as protecting the citizens from the evil other country that’s building AI. “Without strong AI, we can’t build weapons to defend the country.” and “without strong AI, our companies won’t be able to compete in the world marketplace.”
Not serious, not worth reading.
Anyone with anxieties over immigration should have those same concerns over AI, many times over.
Skilled immigrants just got a $100,000/year head tax in the US. Where is such a tax for AI?