• FireRetardant@lemmy.world
    link
    fedilink
    arrow-up
    40
    ·
    5 months ago

    I wonder how much energy google wastes on its AI service in the regular search just to give me a worse answer than the top results I was actually looking for.

    • frezik@midwest.social
      link
      fedilink
      arrow-up
      3
      ·
      5 months ago

      Unlike purchasing things for imaginary gods, carbon credits could work in theory. At least well enough to be part of the solution. That is, if they were properly regulated around strategies that actually absorb carbon and everyone is forced to be honest and transparent.

      Which none of them do, of course.

  • pieter91@sh.itjust.works
    link
    fedilink
    arrow-up
    11
    ·
    5 months ago

    The fundamental problem is that there’s money to be made by consuming more and more “sustainable” resources. The real solution is to reduce consumption on a global scale.

      • greyw0lv@lemmy.ml
        link
        fedilink
        arrow-up
        8
        arrow-down
        1
        ·
        5 months ago

        Not OC, but some ways to “reduce consumption” are reducing our usage of inefficient technology by replacing it with more energy/resource efficient means.

        Examples include replacing individual automobiles with mass transit, building more dense cities to reduce consumption of construction materials/ vehicle miles, and not training massively large language models in facilities that consume more energy than an entire small country.

        • Lotarion@lemmy.world
          link
          fedilink
          arrow-up
          3
          ·
          5 months ago

          … gotta admit this is quite a bit more sound than I anticipated

          As for LLMs, people don’t really like when others say they can’t explore the applications of tech, even if it’s unsustainable, so there’f bacaklash ofc

        • NomenCumLitteris@lemmy.ml
          link
          fedilink
          arrow-up
          3
          ·
          5 months ago

          In real world application, increased efficiency doesn’t decrease energy usage nor decrease labor required to live. Tech has gotten more efficient since the industrial revolution, but demand for technology has increased exponentially, energy use is astronomical, and workers still work more hours.

    • jsomae@lemmy.ml
      link
      fedilink
      arrow-up
      2
      ·
      5 months ago

      This may be true of chopping down forests or mining coal. But we can use nuclear power. And the earth has plenty of water – does chatgpt need clean drinking water specifically?

      • frezik@midwest.social
        link
        fedilink
        arrow-up
        4
        ·
        5 months ago

        Datacenters moved to using evaporative cooling to save power. Which it does, but at the cost of water usage.

        Using salt water, or anything significantly contaminated like grey water, would mean sediment gets left behind that has to be cleaned up at greater cost. So yes, they generally do compete with drinking water sources.

        There’s no way nuclear gets built out in less than 10 years.

        • jsomae@lemmy.ml
          link
          fedilink
          arrow-up
          1
          ·
          5 months ago

          Thank you for explaining that. I didn’t understand the need to use drinking water.

    • TheOubliette@lemmy.ml
      link
      fedilink
      arrow-up
      1
      ·
      5 months ago

      Flip it around a little: we need to take control over production to eliminate this pointless and even pernicious waste. There is such massive waste in this system, so much energy and resources and lives dedicated to harmful or wasteful activities, that never really touches a consumerist perspective.

      For example, the for-profit healthcare insurance system in the US. If you fired 90% of them, ran a central insurance option through the government, and then paid every single person you fired the same just to do whatever they wanted, you’d actually make the system better and more efficient. That 90% are not just redundant, they are there to put up barriers for needed healthcare because that makes the company money.

      The more you analyze any industry, the more you will find these attributes. The product that doesn’t need to exist and only does because of some other deficiency driven by capitalism. The massive bureaucies dedicated to monitoring workers so they don’t unionize, the massive bureaucies that must be duplicated across 50 companies because they each have to do accounting and taxes and payroll and answer calls even though they make the same widget.

      On top of all of this is war and related imperialism. Entire countries are thrown into chaos, with this economic system as the root cause. Why is Venezuela so heavily sanctioned? Simply because they nationalized their oil industry so that the money made would go to Venezuelans. This ran against the capitalist imperialist ownership of Venezuela’s resources so they did their very best to destroy the country using economic means. Think of all the people forced into poverty because of this. Think of what they could have built instead. Now think of Iraq, its infrastructure bombed to nothing. We should center the people, but also think of all of the resources it takes to build a power grid, a clean water system, desalinization plants, roads, etc. All of that rubble because Iraq stepped outside the domination of US Empire, itself just an extension of global capital.

      Through this we will decrease consumption as well, it is the natural outcome of not maintaining so much redundancy, of destroying so much of what is built, of being able to focus on real problems and developing real solutions rather than forcing humanity into pointless tasks.

  • Kachajal@lemmy.ml
    link
    fedilink
    arrow-up
    10
    arrow-down
    7
    ·
    5 months ago

    Ehhh. I get that exploitative techbros and cryptobros have confused the issue by latching onto the AI bubble.

    But at the same time generalized artificial intelligence is very likely possible and will be an absolute game-changer if and when it happens. It’s easily of similar value to fusion technology.

    And it is already bringing truly impressive results into reality - protein folding and diagnostic medicine come to mind.

    • abbiistabbii@lemmy.blahaj.zone
      link
      fedilink
      arrow-up
      17
      arrow-down
      1
      ·
      5 months ago

      But at the same time generalized artificial intelligence is very likely possible and will be an absolute game-changer if and when it happens. It’s easily of similar value to fusion technology.

      The “AI” we have now is basically advanced Autocomplete.

      • Kachajal@lemmy.ml
        link
        fedilink
        arrow-up
        9
        arrow-down
        5
        ·
        5 months ago

        In the same way that computers are basically advanced abacii.

        Don’t confuse a simplification made to demonstrate the basic functioning to a layman with how things actually work.

        LLM’s are neural networks, which are based on a model of brain function. There’s little reason to believe that we cannot eventually reach similar levels of effectiveness as human brains.

        Hell - reaching the levels of pigeon brains would already be absurdly useful.

        • chayleaf@lemmy.ml
          link
          fedilink
          arrow-up
          7
          ·
          5 months ago

          While I agree that LLMs can achieve human-tier efficiency at most tasks eventually (some architectural changes will be necessary, but the core approach seems sound), it’s wrong to say it’s modeled after the human brain. We have no idea how brains work as they’re super complex, we’re building artificial neural networks from the ground up. AI uses centuries’ worth of math, but with our current maths knowledge the code isn’t too complicated. Human brains aren’t like that, they can’t be summed up in a few lines of code because DNA is a huge mess that contains so much more than just “learning”, so many inactive or redundant bits and pieces. We’re building LLMs with knowledge of how languages work, not how brains work.

          • jsomae@lemmy.ml
            link
            fedilink
            arrow-up
            0
            ·
            5 months ago

            Transformers are not built with our knowledge of language. That’s a gross approximation – it would honestly be more accurate to say they’re modelled after the human brain than that they’re built with our understanding of language. A big problem is that the connection between AI and language is poorly understood – we can’t even understand what the word2vec axes are.

            • chayleaf@lemmy.ml
              link
              fedilink
              arrow-up
              1
              ·
              5 months ago

              i’m not talking about knowing about how humans perceive/learn languages, i’m talking about language structure. Perhaps it’s wrong to call it “how languages work”

              • jsomae@lemmy.ml
                link
                fedilink
                arrow-up
                1
                ·
                5 months ago

                That’s what I meant, yes. They’re not built based on any linguistic field

                • chayleaf@lemmy.ml
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  edit-2
                  5 months ago

                  different neural network types excel at different tasks - image recognition was invented way before LLMs, not only for lack of processing power, but also because the previous architectures didn’t work with languages. New architectures don’t appear out of thin air, they are created with a rough idea of what we could need to make the network do a certain task (e.g. NLP) better. Even tokenization isn’t blind codepoint separation but is based on an analysis of languages. But yes, natural languages aren’t “parsed” for neural networks, they don’t even have a formal grammar.

        • pelotron@midwest.socialOP
          link
          fedilink
          English
          arrow-up
          6
          ·
          5 months ago

          The problem is they’re already talking about needing trillions of dollars worth of hardware to make it happen. It’s absurd.

  • Raphaël A. Costeau@lemmy.ml
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    7
    ·
    edit-2
    5 months ago

    Transition to paperless office

    The problem is there is people who say bullshit like this unironically.

    • Lotarion@lemmy.world
      link
      fedilink
      arrow-up
      3
      ·
      5 months ago

      And it doesn’t even make sense

      99% of a modern office’s correspondence already goes on online, and only the most important stuff gets backed up on paper copies, often because of regulations that are there for a reason

    • dislocate_expansion@reddthat.comB
      link
      fedilink
      arrow-up
      3
      ·
      5 months ago

      Seriously. It isn’t helpful towards the environment if we are using so many resources to mine chips and metals and then push it along the internet to then be trained on said AI bots. Would be more sustainable using paper and planting trees smh