Goats in the Machine 康友泥
ricebucket1.bsky.social
Goats in the Machine 康友泥
@ricebucket1.bsky.social
they/them/X也 | agropunk | (political) ecologist | remote-sensing/earth-observation | counter-mapping | lapsed anthropologist | Middle Rio Grande | Taiwanese-American | more than human ethnography |
This is something I believe personally, but I'm not sure it's effective strategy. Pretty much everyone considers plants and animals alive, but that hasn't stopped wanton habitat destruction.

Maybe it's a harder sell, but I think the non-human persons concept carries more weight.
May 28, 2025 at 1:57 AM
But I do hear people say things (on the internet, or while traveling) about how water is renewable because of the water cycle - so why conserve it? I wonder if that's a feature of living in moist climates where water stress is less visible.
May 19, 2025 at 3:03 PM
At least here in NM average people are generally aware that water is a limited resource. Some places have had good success significantly reducing per capita water consumption (e.g. Las Vegas). It's hard to ignore the fact that you're living in a desert.
May 19, 2025 at 3:01 PM
I'm all about the dunking on Elon, but I don't think it should be done based on a single data point from a random forum post.
May 19, 2025 at 1:41 AM
So for some reason I looked at the source (cybertruck owner forum). Gross yet fascinating. But I did also notice that someone else had a resale loss of $39k with 32,000 miles. So it clearly doesn't scale linearly (i.e. 5.6 cents/mi). The other person would have seen a loss of 1.2 cents/mi.
May 19, 2025 at 1:40 AM
I'm pretty familiar with water issues in the western US, from living here and working in water resources--but I was surprised to see how much of the northeast was considered high or very high water stress. I grew up in NJ and now live in New Mexico - both areas were in the same water stress class.
May 18, 2025 at 4:39 PM
Looks amazing! Do you know if they are going to do more performances after this weekend? I'm in the (very) early stages of my own work on railroads and Chinese diaspora in the southwest and could always use more inspiration :)
May 16, 2025 at 5:37 PM
Yeah I did some back of the envelope calculations and I think several hundred queries consumed the same amount of water as a single almond.

However if we all stopped querying, maybe it would stop or slow the intensity of model training.
May 8, 2025 at 9:47 PM
Obviously it'd be better if they were being built in low water stress areas (or not at all), but that's less than 9% of the area in the US. It's really just northern Maine, some of the Washington coast, the north third of Missouri and some of the Dakotas.
May 8, 2025 at 9:44 PM
Really great article, but one piece of context I feel is important to mention is that ~ 2/3 of the contiguous US is categorized as high or extremely high water stress (according to WRI). So the fact that 2/3 of the data centers are in those areas is not necessarily out of proportion.
May 8, 2025 at 9:32 PM
I'm not sure how that actually translates to strategy, but I do think there's a possibility that discussing the "plight" of young men could make sense within a harm reduction framework.
May 4, 2025 at 8:24 PM
When this has come up in conversation in my community - it's generally around the idea that young men believing they are oppressed/ignored - while still holding hegemonic power in practice - leads to them inflicting violence on actually marginalized groups.
May 4, 2025 at 8:22 PM
It'd be like asking a bike mechanic to oversee the development of a municipal public transit system.

I'd much rather see oversight come from Science and Technology Studies, anthropologists and ethics people than scientists.
April 27, 2025 at 8:31 PM
One issue with scientific oversight is that the vast majority of ML models are built to serve a very specific purpose (e.g. IDing pronghorn using wildlife crossings from trail cam photos). A wildlife ecologist isn't really qualified to oversee massive general purpose LLMs.
April 27, 2025 at 8:29 PM
Also I'm sorry if my neuro-divergent brain is hyper fixating on your use of the word "nature". I think we're generally in agreement about abhorrent promotion of AI by tech companies.

My daily exposure to AI is mainly folks using it for ecological research, so that's my bias.
April 27, 2025 at 8:09 PM
There's lots of important work in the sciences being done by amazing folks, and I don't think it's fair to say that anyone doing any kind of ML work is a thief.

I don't think that's what you meant, but that's what saying ML is theft by nature implies.
April 27, 2025 at 7:56 PM
Yes of course stealing copyrighted material is theft.

But I don't think me going into a forest and identifying tree species is theft.

If it's possible to ethically collect training data, even if it isn't often done, than it's not theft "by nature".
April 27, 2025 at 7:53 PM
Totally agree! And I realized I don't actually know how "generative AI" is defined, so maybe most of the ML/deep-learning tools being used in the sciences actually don't count.
April 27, 2025 at 7:47 PM
I wasn't objecting to the general awfulness of generative AI, just to the statement that the creation of training data is stealing by nature.
April 27, 2025 at 7:28 PM
I'm hella skeptical of genAI for all sorts of systemic reasons, but it definitely "works" in certain cases. In the computer vision for ecology world, confusion matrices are a fairly objective way of assessing the accuracy of a model. We can't use models that aren't accurate.
April 27, 2025 at 7:20 PM
As someone who uses ML quite a bit for work, stealing/scraping data is not inherent to the process. I work in habitat restoration, and we collect training data in the field in order to classify water surface extents, tree species and vegetation health.
April 27, 2025 at 7:15 PM
I've had it work well in a huge variety of use cases. Sure it requires lots of vetting, and you need to already be somewhat proficient in the language. But that's also true with copy and pasting random answers from stackexchange, which is generally what happened before LLM coding assistance.
April 27, 2025 at 7:06 PM
My (very limited) understanding is that the vast majority of the resource use is from training the models. So I think the path of least damage would be to use the already trained model, instead of training your own new one. Assuming they need to choose one of those options.
April 27, 2025 at 7:03 PM
I agree on this one - it's helped me troubleshoot programming issues that I never could have figured out otherwise (no documentation, only thing I could google were a few stackexchange questions about the same package but totally unrelated issue)
April 27, 2025 at 7:00 PM