tk,
@tk@bbs.kawa-kun.com avatar

:blobclown:

newt,
@newt@stereophonic.space avatar

@tk what's nightshade?

icedquinn,
@icedquinn@blob.cat avatar

@newt @tk i think there was a project to try and watermark images in ways that could make common diffuser models confused if you tried to train art on it. i kind of just laughed at the concept because the horny nerds will find a way around it.

TURBORETARD9000,

@icedquinn @tk @newt It's a little more advanced than watermarking but yeah
It's fairly effective from what I've seen
They're just straight up poisoning the well from which training data comes by tweaking pixels in a way that's barely perceptible to humans (still working on that) but deadly to training

Moon,
@Moon@shitposter.club avatar

@TURBORETARD9000 @tk @icedquinn @newt I suspect it will remain effective since you can trip up AI pretty easily in lots of other ways

TURBORETARD9000,

@Moon @tk @icedquinn @newt yeah
I don't think we'll see any winners any time soon

Moon,
@Moon@shitposter.club avatar

@TURBORETARD9000 @tk @icedquinn @newt I think big models are doomed, the future is curated models grown over time.

leyonhjelm,
@leyonhjelm@breastmilk.club avatar

@Moon

Unlicensed models with illegal datasets are what will be used, with something else on the public face

@TURBORETARD9000 @tk @icedquinn @newt

Moon,
@Moon@shitposter.club avatar

@leyonhjelm @TURBORETARD9000 @tk @icedquinn @newt yeah there will be models that didn't ask permission but they'll be hand-curated so they'll avoid poisoning

TURBORETARD9000,

@Moon @tk @icedquinn @leyonhjelm @newt > hand-curated
Lol, lmao
That's not possible except for the truly dedicated
You need millions of samples, each which would need to be checked individually

Moon,
@Moon@shitposter.club avatar

@TURBORETARD9000 @tk @icedquinn @leyonhjelm @newt start with the existing non-poisoned smart guy

Moon,
@Moon@shitposter.club avatar

@TURBORETARD9000 @icedquinn @leyonhjelm @newt @tk also for images it takes waaaay less data than you think, you can make your own lora with like 40 images

TURBORETARD9000,

@Moon @tk @icedquinn @leyonhjelm @newt wh
I thought you needed at least 10k samples just to do simple image identification tasks

icedquinn,
@icedquinn@blob.cat avatar

@TURBORETARD9000 the amount of data you need to train from random init depends on how large the network is. you do need a "lot" of images when you make some huge chungus neural network.

fine tuning you can get away with less. people have done fine tunes of VITS speech models with only minutes of samples.

the trick is all the other data that was close-ish is still in the neural network.

that being said existing models are severely overblown. nobody us doing stuff like the old GMDH where you grow a network that is just the right size, or Numenta's contextualized learning, they're just shitting tons of neurons out and boiling oceans.

stanford already proved most of the ML model is horseshit https://dawn.cs.stanford.edu/2019/06/13/butterfly/

@tk @leyonhjelm @Moon @newt

Moon,
@Moon@shitposter.club avatar

@icedquinn @TURBORETARD9000 @tk @leyonhjelm @newt oh I meant for generating "new things". Like if I want to add a single new thing to the model via a lora I can do it with few items

icedquinn,
@icedquinn@blob.cat avatar

@Moon @TURBORETARD9000 @tk @leyonhjelm @newt yeah a lora is a fine tuning encoded as some kind of patch if i recall

icedquinn,
@icedquinn@blob.cat avatar

@Moon @TURBORETARD9000 @tk @newt all the adversarial attacks i've seen in whitepapers require you have the model on hand to deceive it. its entirely possible that just fine tuning the model after the fact breaks your cheats.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • random
  • Hentai
  • doujinshi
  • announcements
  • general
  • All magazines