I am so over hearing about AI. It’s getting to the point that I can assume anyone dropping the term at work is an idiot that hasn’t actually used or utilised it.
It’s this LLM phase. It’s super cool and a big jump in AI, but it’s honestly not that good. It’s a handy tool and one you need to heavily scrutinise beyond basic tasks. Businesses that jumped on it are now seeing the negative effects of thinking it was magic from the future that does everything. The truth is, it’s stupid and people need to learn about it, understand it, and be trained in how to use it before it can be effective. It is a tool, not a solution—at least for now anyways.
I think more like discovering making fire or something. 90% of all the energy burnt is people worshipping it as it blazes away, never actually fulfilling any practical use except being marvelous to be around.
But once the forest is all chopped down, people are forced to understand fire and realise a couple small logs in a contained place was all they needed to have it be incredibly effective.
Oh, but that’s too hard. It’s magic right now. All hail the AI bonfire!
Massive energy consumption. Huge datacenters and not enough green energy. Now they want to build small nuclear plants. Without talking about the waste problem.
Ah yeah, just choose a different energy souce. Simples.
Have you seen the growth in % of renewable (incl, nuc biofuel and waste) electricity generation over the past 30 years.
(36% i in 1990 , dropped to about 33% in late 2000s up to 38% recently) this is global, IEA figures.
There have been two years since 1990 when renewable electricity output has grown faster than total electricity demand.
2008/9 recession and 2020 covid.
The only way renewables will come close to meeting current electricity consumption is actually to start reducing those demands.
If we start transerffing gas( domestic heating), and petrol( low-capacity road transportation) onto the electricitry grid then the scale and speed of renewables needs to ramp up inconcievably quickly - it has grown fast over the past decade, but it hasn’t been cheap nor has it been fast enough to keep up with current demands.
TBH I don’t know where AI lines up next to EVs in scale of potential extra demand, probably lower but still an added demand (unless it can substitute for other stuff and improve efficiency somehow).
Electricity source is not really a choice, it is resource and tech constrained many sources are needed; the cheapest fuels will continue to be in the mix used so long as demand keeps increasing so fast.
Maybe, If you ran all AI in peooles houses in cold countries in winter, it’d substitute for heating - that’d be one way it could reduce its impact. Or maybe it can get its act together and spark widespread, frequent, deep, long lasting recessions in economic activity.
Maybe renewables is not the solution to our energy needs if it can’t scale up like we thought it could. Conservation of energy is not the answer. We as a society must find new, cleaner, sources of energy. Maybe AI can help us do it.
Perpetual motion machines are one of the mothers of all snake oil.
Maybe AI can turn base metals into gold too.
Do these AIs even really have a demonstrable understanding of thermodynamics yet?
It needs to prove itself with a usable output for a clear observabe application on a small scale scale before anyone should start chucking vast amounts of energy at it in hope of what it can “maybe” do.
I’d much rather chuck all that energy into trials of tokamaks or something like that.
Their waste is less destructive than coal plant though. Perhaps this could be a silver lining to finally get nuclear back in action and get closer to dropping coal once and for all.
Tbf the energy issues are getting better, or at least there are some more efficient models being created. Back in April there was a version of Llama that only needed 8gb to almost match GPT4
I equate an AI to an intern. It’s useful for some stuff but if I’m going to attach my name to it I’m going to review it and probably change a lot about it.
There’s one good use case for me: produce a bigload of trialcontent in no time for load testing new stuff. “Make 2000 yada yada with column x and z …”. Keeps testing fun and varied while lots of testdata and that it’s all nonsense doesn’t matter.
I’ve found that testing code or formulas with LLM is a 50/50 now. Very often replying “use function blabla() and such snd so” very detailed instructions while this suggested function just doesn’t exist at all in certain language asked for… it’s still something I’ld try if I’m very stuck tho, never know.
Very often replying “use function blabla() and such snd so” very detailed instructions while this suggested function just doesn’t exist at all in certain language asked fo
I’ve noticed this a lot too—especially for M. But even though it makes up a function, it sometimes inspires a more optimised idea/method that can be more flexible for future datasets.
But most times it starts to massacre things and disregard prompted parameters or even producing an identical suggestion immediately after being told not to, why not to, and reconfirming original parameters of the query.
Some times punching in the same prompts five times for five iterations produces completely different results, but one may be on the right track and I can code the rest. It helps to set it’s personality first, so it’s sharing ideas it’s seen out there, rather than trying to please.
At the least, it’s a big time saver. Gone are the days where I get a few days spare to work on solving a complex problem through trial and discovery, so it’s an excellent tool for reducing testing time and speeding up the route to an optimised method.
I am so over hearing about AI. It’s getting to the point that I can assume anyone dropping the term at work is an idiot that hasn’t actually used or utilised it.
It’s this LLM phase. It’s super cool and a big jump in AI, but it’s honestly not that good. It’s a handy tool and one you need to heavily scrutinise beyond basic tasks. Businesses that jumped on it are now seeing the negative effects of thinking it was magic from the future that does everything. The truth is, it’s stupid and people need to learn about it, understand it, and be trained in how to use it before it can be effective. It is a tool, not a solution—at least for now anyways.
So, like a hammer. A very expensive, environment-destroying hammer.
That’s actually a pretty good analogy.
I think more like discovering making fire or something. 90% of all the energy burnt is people worshipping it as it blazes away, never actually fulfilling any practical use except being marvelous to be around.
But once the forest is all chopped down, people are forced to understand fire and realise a couple small logs in a contained place was all they needed to have it be incredibly effective.
Oh, but that’s too hard. It’s magic right now. All hail the AI bonfire!
Genuinely curious, how does it destroy the environment?
Massive energy consumption. Huge datacenters and not enough green energy. Now they want to build small nuclear plants. Without talking about the waste problem.
So AI uses energy, and it’s how we are choosing to provide that energy is destructive to the environment? So AI isn’t itself destructive.
Ah yeah, just choose a different energy souce. Simples.
Have you seen the growth in % of renewable (incl, nuc biofuel and waste) electricity generation over the past 30 years. (36% i in 1990 , dropped to about 33% in late 2000s up to 38% recently) this is global, IEA figures.
There have been two years since 1990 when renewable electricity output has grown faster than total electricity demand. 2008/9 recession and 2020 covid. The only way renewables will come close to meeting current electricity consumption is actually to start reducing those demands.
If we start transerffing gas( domestic heating), and petrol( low-capacity road transportation) onto the electricitry grid then the scale and speed of renewables needs to ramp up inconcievably quickly - it has grown fast over the past decade, but it hasn’t been cheap nor has it been fast enough to keep up with current demands.
TBH I don’t know where AI lines up next to EVs in scale of potential extra demand, probably lower but still an added demand (unless it can substitute for other stuff and improve efficiency somehow).
Electricity source is not really a choice, it is resource and tech constrained many sources are needed; the cheapest fuels will continue to be in the mix used so long as demand keeps increasing so fast.
Maybe, If you ran all AI in peooles houses in cold countries in winter, it’d substitute for heating - that’d be one way it could reduce its impact. Or maybe it can get its act together and spark widespread, frequent, deep, long lasting recessions in economic activity.
Maybe renewables is not the solution to our energy needs if it can’t scale up like we thought it could. Conservation of energy is not the answer. We as a society must find new, cleaner, sources of energy. Maybe AI can help us do it.
Perpetual motion machines are one of the mothers of all snake oil. Maybe AI can turn base metals into gold too. Do these AIs even really have a demonstrable understanding of thermodynamics yet? It needs to prove itself with a usable output for a clear observabe application on a small scale scale before anyone should start chucking vast amounts of energy at it in hope of what it can “maybe” do. I’d much rather chuck all that energy into trials of tokamaks or something like that.
Their waste is less destructive than coal plant though. Perhaps this could be a silver lining to finally get nuclear back in action and get closer to dropping coal once and for all.
Tbf the energy issues are getting better, or at least there are some more efficient models being created. Back in April there was a version of Llama that only needed 8gb to almost match GPT4
I equate an AI to an intern. It’s useful for some stuff but if I’m going to attach my name to it I’m going to review it and probably change a lot about it.
There’s one good use case for me: produce a bigload of trialcontent in no time for load testing new stuff. “Make 2000 yada yada with column x and z …”. Keeps testing fun and varied while lots of testdata and that it’s all nonsense doesn’t matter.
I’ve found that testing code or formulas with LLM is a 50/50 now. Very often replying “use function blabla() and such snd so” very detailed instructions while this suggested function just doesn’t exist at all in certain language asked for… it’s still something I’ld try if I’m very stuck tho, never know.
I’ve noticed this a lot too—especially for M. But even though it makes up a function, it sometimes inspires a more optimised idea/method that can be more flexible for future datasets.
But most times it starts to massacre things and disregard prompted parameters or even producing an identical suggestion immediately after being told not to, why not to, and reconfirming original parameters of the query.
Some times punching in the same prompts five times for five iterations produces completely different results, but one may be on the right track and I can code the rest. It helps to set it’s personality first, so it’s sharing ideas it’s seen out there, rather than trying to please.
At the least, it’s a big time saver. Gone are the days where I get a few days spare to work on solving a complex problem through trial and discovery, so it’s an excellent tool for reducing testing time and speeding up the route to an optimised method.