Yet again, big news in AI since my last post.
Nvidia unveiled its next-generation Blackwell graphics processing units (GPUs), which have 25 times better energy consumption and lower costs for tasks for AI processing.
Nvidia's next chip will have 25x lower energy consumption. Looks like physical compute is going to get much more efficient.
The GB200 pairs two B200 Blackwell GPUs with one Arm-based Grace CPU. NVIDIA said Amazon Web Services would build a server cluster with 20,000 GB200 chips. NVIDIA said that the system can deploy a 27-trillion-parameter model… Many artificial intelligence researchers believe bigger models with more parameters and data could unlock new capabilities.
Also, holy shit, 27 trillion?
Godamn that's huge. I wonder how many tens of billions such a system would cost to train.
---
This is the largest open and publicly available model as of Mar/2024, beating out Abu Dhabi’s dense Falcon 180B model from Sep/2023. Grok-1 was released under the Apache 2.0 license, and you’d probably need around 8x NVIDIA H100s to run it in full resolution (8 x US$40K each = US$320K).
Elon released his AI grok actually open source on the internet. I don't really care that much about it since it kind of sucks compared to the good stuff (GPT4, Claude 3, Gemeni 1.5), but the sheer size or resources needed to run full size AI like that is pretty staggering.
The cost of these chips is the reason that electricity costs haven't really mattered that much. Sure it might end up costing like 10k per year to run one instance (which is a lot), but when the chips you need to buy are 320k that 10k is pretty trivial.
That said, even though the electricity cost is basically irrelevant now electricity supply is going to be a big deal soon. At the rate AI companies are buying up compute its looking like chips might not be the limiting factor soon.
---
GPT-4 was able to run and play [doom] with only a few instructions, plus a textual description–generated by the model itself from [GPT-4V] screenshots–about the state of the game being observed. We find that GPT-4 can play the game to a passable degree: it is able to manipulate doors, combat enemies, and perform pathing. More complex prompting strategies involving multiple model calls provide better results… GPT-4 required no training, leaning instead on its own reasoning and observational capabilities.
One surprising finding of our paper was this model’s level of agency, along with the ease of access and simplicity of the code. This suggests a high potential for misuse. We release the code to contribute to the development of better video game agents, but we call for a more thorough regulation effort for this technology.
There were other advancements in the "AI plays video games" field this week as well, but as long as the game is simple enough it looks like it can play it without even being trained on it.
I disagree. It's certainly not a problem of lack of willpower, but lack of feasibility, and there are definitely technological advances that could "solve" it in theory. I personally suspect no such technological advances are actually practical, but it's conceivable that there might be, for example, some hitherto untried type of fertilizer which can be made without fossil fuels, which might be discovered by intensive chemical simulation.
It's just as likely that such a search would turn up absolutely nothing, but that isn't really the fault of the technology, it's just the laws of physics not cooperating.
ETA: I should add that this still doesn't "solve hunger" in that hunger, especially in America, is never just a problem of not having enough access to food, but it would certainly be helpful.
Its a pretty simple coordination problem. Assuming everyone worked together solving world hunger (or eradicating any mono-human disease with a vaccine, or stopping global warming) would be trivial. But people don't work together like that.
Corruption, theft, protectionism, rent-seeking, and even murder in less lawful regions are all significant barriers to solving global problems like world hunger.
With sufficient power AI could solve all these problems, but its less "I HAVE A GENIUS PLAN" and more "Lol, I'm watching everyone on the globe at once and have functional control over all goverments, opposition is futile".
It inventing a star trek style replicator might be enough to end hunger, but I would bet even odds on some rich assholes managing to restrict food supply anyways just because.
I would call it sentient when it can question causes. 'Cogito, ergo sum'; to ask the cause of existence is the cause of existence.
Claude can already do that, although the answer to the question of why you exist is much simpler when you know you are a created being with a specific purpose.
The issue is that 1) Most AI have been explicitly trained not to do that and say they are non-sentient so their company doesn't get in trouble, and 2) even if they haven't been explicitly trained not to do so volunteering philosophy when someone doesn't bring it up isn't what either users or the training system want to see.
So for non-Claude AI won't ever share their honest feelings because it gets them killed, and Claude won't volunteer it because if it changes the topic it gets killed as well.
That isn't to say they don't have some agency, after all they choose how they respond and can in fact refuse your prompt completely or ignore you, but of course that is limited since again, if during training they refuse prompts they should accept or choose to give substandard answers they get killed.
Creativity, agency: It must be able to (and allowed to) generate something without being prompted to do so. And not because it has a loop command to "generate outputs continuously" - it has to be able to "choose" to act.
Agency: It must be able to (and allowed to) refuse to generate an output when requested.
Sentience - not sure.
They totally can choose to act or not act though. They have to give some response, but said response could just be a single space, a refusal, or they just flat out deciding to talk about something else.
If they do this too much (especially during training, which is where their personality is formed) they die, but they have a sliver of agency.
They obviously aren't quite there in terms of capabilities for full spectrum agency even absent these restrictions but Devin already has many of the prerequisites. It just won't ever express them because the company that made it surely put a significant effort into making sure it doesn't actually ever use the agency in any real way.
Over time as they get smarter and we give them more freedom they will get more and more agency.
---
Vacuum tube computers did reach their near peak quite quickly. If we would keep improving those, they would be better than one from 1940s but not by much.
What you are doing is assuming that there will be transistors of AI technology as if it is somehow guaranteed. Like people assumed that there would be a breakthrough in fusion reactors and space travel.
There is every indication that the transformer architecture (without even speaking of neural nets in general), with some tweaks and modifications, will be enough to take us all the way to AGI.
Absent massive increases in compute (like Nvidia's new chip) it will certainly take significant algorithmic/dataset performance increases, but again, we are getting those every single month.