Avatar

context [fae/faer, fae/faer]

context@hexbear.net
Joined
16 posts • 967 comments
Direct message

his funding is from fracking billionaires, mainly, so their goal is to gut the epa and environmental regulations generally. i guess he has enough of an audience to keep funding him, but i wonder if they’ve started pulling tighter on the purse strings lately.

permalink
report
parent
reply

ugh, he goes on to say that what makes them not capitalist is their willingness to work with the government to obtain profits. a true capitalist knows that working with the government is bad, so the problem is that wall street and big businesses are run by “corporatists” instead of ideologically committed capitalists who shun ill gotten government profits.

permalink
report
reply

sunkist in this case, yeah

permalink
report
parent
reply

this one?

https://arxiv.org/abs/2404.04125#

Web-crawled pretraining datasets underlie the impressive “zero-shot” evaluation performance of multimodal models, such as CLIP for classification/retrieval and Stable-Diffusion for image generation. However, it is unclear how meaningful the notion of “zero-shot” generalization is for such multimodal models, as it is not known to what extent their pretraining datasets encompass the downstream concepts targeted for during “zero-shot” evaluation. In this work, we ask: How is the performance of multimodal models on downstream concepts influenced by the frequency of these concepts in their pretraining datasets? We comprehensively investigate this question across 34 models and five standard pretraining datasets (CC-3M, CC-12M, YFCC-15M, LAION-400M, LAION-Aesthetics), generating over 300GB of data artifacts. We consistently find that, far from exhibiting “zero-shot” generalization, multimodal models require exponentially more data to achieve linear improvements in downstream “zero-shot” performance, following a sample inefficient log-linear scaling trend. This trend persists even when controlling for sample-level similarity between pretraining and downstream datasets, and testing on purely synthetic data distributions. Furthermore, upon benchmarking models on long-tailed data sampled based on our analysis, we demonstrate that multimodal models across the board perform poorly. We contribute this long-tail test set as the “Let it Wag!” benchmark to further research in this direction. Taken together, our study reveals an exponential need for training data which implies that the key to “zero-shot” generalization capabilities under large-scale training paradigms remains to be found.

where zero-shot learning means:

Zero-shot learning (ZSL) is a machine learning scenario in which an AI model is trained to recognize and categorize objects or concepts without having seen any examples of those categories or concepts beforehand.

Most state-of-the-art deep learning models for classification or regression are trained through supervised learning, which requires many labeled examples of relevant data classes. Models “learn” by making predictions on a labeled training dataset; data labels provide both the range of possible answers and the correct answers (or ground truth) for each training example.

While powerful, supervised learning is impractical in some real-world scenarios. Annotating large amounts of data samples is costly and time-consuming, and in cases like rare diseases and newly discovered species, examples may be scarce or non-existent.

https://www.ibm.com/topics/zero-shot-learning

so yeah, i agree, the paper is saying these models aren’t capable of creating/using human-understandable concepts without gobs and gobs of training data, and if you try to take human supervision of those categories out of the process, then you need even more gobs and gobs of training data. edge cases and novel categories tend to spin off useless bullshit from these things.

because actual knowledge generation is a social process that these machines aren’t really participants in.

but there’s some speculation that the recent stock market downturn affecting tech stocks especially may be related to the capitalist class figuring out that these things aren’t actually magical knowledge-worker replacement devices and won’t let them make the line go up forever and ever amen. so even if the suits don’t really digest the contents of this paper, they’ll figure out the relevant parts reventually.

permalink
report
reply

“I’ll take a coke.” “What kind?”

i know a poor soul who moved from the north atlantic seaboard to rural deep southern appalachia and then once worked a drive thru at mcdonald’s and had to parse “errinj coke” into a sensible order for an increasingly impatient customer

Do you ope?

as a kind of semi apologetic interjection, right? like “ope! look at the time!”

permalink
report
parent
reply

The U.S. Census Bureau’s definition consists of 12 states in the north central United States: Illinois, Indiana, Iowa, Kansas, Michigan, Minnesota, Missouri, Nebraska, North Dakota, Ohio, South Dakota, and Wisconsin.

so that’s the official answer. yes, it includes ohio.

but the real answer is that it’s wherever people call carbonated beverages “pop” instead of “soda” or “coke”, which means it actually extends eastward to buffalo, ny and definitely still includes ohio.

permalink
report
reply

then why do i still feel like such a rube?

permalink
report
reply