No.843[View All]
Anyone else been messing around with the stable diffusion algorithm or anything in a similar vein?
It's a bit hard to make it do exactly what you want but if you're extremely descriptive in the prompt or just use a couple words it gives some pretty good results. It seems to struggle a lot with appendages but faces come out surprisingly well most of the time.
Aside from having a 3070 i just followed this guide I found on /g/
https://rentry.org/voldy to get things setup and it was pretty painless.
310 posts and 172 image replies omitted. Click reply to view. No.1154
>>1153I guess if you want to summarize all of this into one issue you could say that Google (Alphabet) has a management issue.
No.1155
>>1150>Holy cow that's a lot of citations.Yeah... This is a bit off-topic: I've mentioned them before on Kissu, but I really recommend the YouTube channel Level1Techs. All of those articles were sourced from previous episodes of their podcast. Thankfully, they source every article they talk about in the description so it was easy to search for keywords and find them. They do a really good job aggregating the news of the week, and go over business, government, social, "nonsense", and AI/robot articles as they relate to tech. The podcast and reviews they do are just something they do on the side, mostly for fun. They run a business that does contracted software/website development so they're very well versed in corporate affairs and the workings of all sorts of tech stuff and I largely trust their opinions on various topics. Naturally, they talk about political things with some regularity, but they're fairly diverse in terms of viewpoints with some disagreement between each other so there's never really any strong political lean to the things they discuss.
No.1156
>>1152From [4]
>When AI fails, it doesn’t fail for [] white menQuite ironic, in retrospect.
>those articles only add evidence of intentionality.I think they do the opposite.
The articles repeatedly present the administration of google as being anti-woke, so to speak, hiring rightwingers for their AI research team, firing leftwingers and censoring papers that criticize its own products for being discriminatory.
After beheading their ethics team, the doubling of the team's size feels like a marketing stunt gone out of control.
No.1157
>>1153Well, as somebody that works on a lot of open source projects, this explains why Google, even when they pretty much take over a project seem to 'lose interest' and stop contributing. I deeply dislike Google, I probably only detest Oracle and IBM more but I feel kind of bad about some of the posts I've made about flighty Googlers. They didn't lose interest in the new shiny, they left or got fired likely.
No.1158
>>1153It's also, from what I've seen, an unsustainable lifestyle to work there, apparently its very flexible and accommodating but they want very long shifts. It makes sense why people would do it just for a recognizable pad after seeing what it's really like.
Just hearsay, though.
No.1161
>>1159>>1160It's a testament to google's monopoly power that a business strategy like that doesn't just tank the whole company.
No.1162
>>1156what needs to be noted is that the original 2019 ATEAC board was disbanded just four days after [4] was published, so the reactionary guy did get booted out as the protesters wanted:
https://www.bbc.com/news/technology-47825833https://blog.google/technology/ai/external-advisory-council-help-advance-responsible-development-ai/>It's become clear that in the current environment, ATEAC can't function as we wanted. So we’re ending the council and going back to the drawing board. We’ll continue to be responsible in our work on the important issues that AI raises, and will find different ways of getting outside opinions on these topics.not only that, inside of google there appears to be a strong and fostered tradition of criticizing upper management whenever someone disagrees, which has resulted in internal protests that hundreds, thousands, or even twenty thousand workers have taken part in and did receive concessions for it. this article is pretty damn long, but i recommend you read it:
https://archive.is/gOrCXit goes over various things, such as the reasons behind unrestricted entrepeneurship (which precedes the creation of alphabet by at least a decade), being blocked in china, and their attempt at obtaining military contracts for the sake of keeping up with competitors like amazon with its ensuing internal backlash. it presents a picture of an organization where there's a strong divide between execs and regular employees, especially activists, who can go as far as broadcasting a live meeting to a reporter for the sake of sabotaging their return to china. its final section ends with ATEAC's disbanding and how the dismantling of mechanisms for dialogue only heightened tensions between the up and down.
then, during the gebru affair of late 2020-early 2021 there too was a big split over the role of AI [6]:
>Gebru is a superstar of a recent movement in AI research to consider the ethical and societal impacts of the technology.and again hundreds of workers protested, leading to the increase in size of the ethics team a few months later. the head of the team and representative from [9], herself a black woman that expressed problems with exclusion in the industry, spoke of making AI that has a "profoundly positive impact on humanity, and we call that AI for social good." there's a really strong record of activism combined with unparalleled permissiveness, autonomy, to back the idea that yes, this scandalous program is working as intended, regardless of what Pichai may wish. they simply went too far in one direction.
No.1163
>>1162Thanks for the continued feeding of articles. (I have nothing else of value to say)
No.1164
>>1163it was an interesting read (neither do I)
No.1166
>>1165It does not seem to have paid much attention to the reference image, or am I missing something?
No.1167
>>1166Well, I mean I was purposely using a different prompt like "sitting". The little pajama skirt thing is there on two of them and the blanket pattern is there. It attempted to make little stuffed animals in the top left with the little information it had.
It was kind of a bad image to use in regards to her face or hairstyle because it's such a small part of the image.
You shouldn't expect miracles. It's just one image.
No.1168
>>1167I understand the sitting part but the only aspects of the image it seems to have take are the bed sheets and blonde hair.
The hairstyle is wrong in every image as is what she is wearing and I think it should have enough to work with regarding both. The furniture does not match but that is to be more expected. I just thought it would be more accurate with regards to the character.
No.1169
>>1168I think the value is more in the expansion of how prompts are input. An image could be worth more than inputting the prompt directly, and when submitted alongside a text prompt for more detail you can make more with less.
I genned this with the reference image on the left, and just "on side" in the prompt. You don't need to specify every detail explicitly if the image does the bulk for you, but it would be a good idea to still explicitly prompt it for things you want.
No.1170
>>1169I suspect that the more popular Touhous would already be in most image generating AIs' training data.
No.1172
>>1170You're correct. It is which is why their names have so much weight for the token as it just gets the clothing, hair, general proportions, and all that without specification. They are statistically significant in the training data. For example on Danbooru, Touhou is the copyright with the most content under it (840k) with an almost 400k lead on the second place.
The thing is I didn't specify Yakumo Ran or kitsune or any of that in the prompt, the image did all the heavy lifting. The image I posted was an outlier where it got the color of clothing right out of a dozen or so retries because it really wanted to either give her blue robes (likely because the image is blue tone dominant) or a different outfit altogether. Granted there are some details common with her outfit that were added but are not present in the reference image, that being purple sleeve cuffs and talisman moon rune squiggles. With the training data being as it is, those things likely have an extremely high correlation and it put them there because that's what it learned to do.
No.1173
>>1172>The thing is I didn't specify Yakumo Ran or kitsune or any of that in the promptYou don't have to.
People have managed to get art generators to create art strongly resembling popular characters using only very vague descriptions, simply because they feature so prominently in their data sets.
This is why, when you want to demonstrate the capabilities of an AI, you should use obscure characters that the AI is not yet familiar with.
No.1174
yeah like the twins
No.1176
woowwwwwww, nice
No.1177
cute feet btw
No.1178
>>1173It also helps when the character has a unique design, I've made Asa/Yoru pics with AI and even with a lot of tags it sometimes makes Asa look like a generic schoolgirl unless you specify one of the two most popular fan artists of her.
Once you specify Yoru with the scarring tags, it very quickly gets the memo of who it's supposed to be. You didn't sign her petition!
One thing is that I've had trouble having szs characters look like themselves, particularly having issues not making Kafuka and Chiri look like generic anime girls, although that is pretty funny.
I use NovelAI's web service. I know, I know, but I'm fine paying them because it's important to have an AI that is designed to be uncensored, and it really is uncensored, also because I use a computer I rescued from being ewaste from a business. Intel i5-8600T (6) @ 3.700GHz Intel CoffeeLake-S GT2 [UHD Graphics 630] and 8gb of ram. It's not a potato but it certainly is not suited to AI work, which may be a reason to get a strong PC (or buy Kissu a strong GPU for christmas) this year.
>>1175Not bad, the funny part is that I could easily see the dump thing happening in PRAD.
No.1179
>>1178>the funny part is that I could easily see the dump thing happening in PRAD.I can't, what episode plot would involve the twins hanging out in garbage?
No.1180
>>1179Not an episode specifically, I mean the girls have wacky hijinks at the dump and the twins show up
No.1181
>>1180rhythm eats a weird piece of meat at the dump
No.1182
That sounds like a pripara bit, but it works for PR
No.1183
I am looking forward to pripara and kind of annoyed how the experience of watching dear my future and rainbow live is getting into a new group of girls for 50 episodes then they get dropped
No.1184
>>1159>>1160>>1161>>1162This company is more powerful than most governments, by the way. What a world we live in
No.1185
>>1184Even though they get regulated regularly and are consistently seen as incompetent on media...
No.1186
They're not even like Samsung who owns half of South Korea and all the government
No.1187
>give anons the power to make anything with AI
>they make chubby girls and futa
grim
No.1189
>>1105I decided to give Pony another try, or more specifically I checked out a finetune of it called AutismMix and it seems quite impressive. It can even do sex positions! There's still errors that pop up, but like AI in general the reason it works is because your brain is turned off when fapping. The Japanese character recognition is mediocre (Reimu works but Sanae doesn't??) but obviously still far better than my own merges that are like 80% furry just so genitals can be generated. I still find it funny that I knew the furry connection within a few weeks and it took other people over a year to notice it. Furries are powerful.
I really don't know how to prompt for it, but I guess I'll learn eventually. Pic related is what it looks like when I try to prompt Kuon (with other tag assistance) and it completely lacks her beauty and charm of course. Unlike what I previously thought, you can train LORAs with even 8gb of VRAM, so my 12 will allow me to make my Kuon and Aquaplus LORAs again, but I have to do the image cropping all over again because it's 1024x1024 instead of 512x512. Soon...
I'm still going to keep my old models around, not just because of the hundreds of LORAs I have that are not compatible with SDXL, but because I like the general style that my merges have. I may try making SDXL/Pony merges, but I'll see how things go first. It seems to have less options when doing the Supermerge process so that may make it easier.
In other news Stable Diffusion came out today (or will very soon) but like all the other AI stuff I don't have any interest until someone makes a 2D model of it.
No.1190
>>1189>In other news Stable Diffusion came out todayErr Stable Diffusion 3 that is
No.1191
I was looking around 4chan today and happened to stumble upon this thread,
https://boards.4chan.org/a/thread/268171362It got me kind of curious because from what I know most models when you try to edit an image directly tend to alter the base image a little bit into something else instead of perfectly editing the image with a specific modification. From what the OP said in a recent post he's using some sort of subscription, but I know that DallE and Gemini don't work like this so it has to be someone's paid SD offshoot that they've tweaked to work in this way. My question is how would you approach doing this in your own SD model? Via controlnet or something? It seems so odd... Of course there's still plenty of areas where it's making unwanted changes like with changing the style of the bras and aspect ratio or whatnot, but for an advancement in AI modifying only specific details of an image it looks like it's doing pretty good.
No.1192
>>1191Looks like some skilful usage of inpainting.
No.1195
>>1191He's just using NAI probably, although I didn't know they offered such a thing.
It's inpainting, yeah. I don't really do it much, but these days you can use controlnet which would probably be superior to NAI's, although they might have their own model for it or something.
You go to the Img2Img tab and then the Inpaint tab, provide the prompt, and ideally set up controlnet. You use 'inpaint_global_harmon
ious" and set denoise to 1 instead of the usual 0.3 or whatever.
No.1762
>>1761Yeah, it's funny how long AI stayed in that relatively
Coming Soon™ state for over a decade until one day it all just blew up and insane progress occurred nearly overnight.
No.2213
https://x.com/STOP_THINKINGAA Jesus Christ...
At some point you have to wonder how much morals weigh out against just being a scumbag and taking full advantage of ai, this guy's making 2.6k a month doing this. And looking at his 18+ version of the account,
https://x.com/HachimidesR18, it's easily visible how it's completely automated. No thought or care put into anything just the same exact poses with a pallet swap. I don't understand why anyone pays for this when they could easily just start up SD on their own and do the exact same thing.
No.2215
>>2213I don't know how people could possibly pay for AI stuff. I will say at least that he must spend a lot of time cleaning up the images. It lacks creativity or personal input so it's the most inane stuff you could imagine and it's why I never do it. It's not quite manual labor, but it's very dull and repetitive.
The part that bothers me is that an AI generator person's "style" is just a combination of real artist tags, which adds insult to injury when people expect (and do) get paid for it. My opinion is that AI really shouldn't be used for anything other than a personal masturbation aid, which it is very
VERY good at. Well, and making throwaway images that people look at for a few seconds to laugh.
The only paid AI stuff that could possibly make sense to me is the extremely niche stuff that has little no real art.
No.2218
>>2213It's good because it's novel
No.2219
>>2213People are dumb and easy to take advantage of. I'd be a multi-millionaire if I didn't have good morals. The internet is designed now so that these people rise to the top. It's why youtube is filled with dumb people making thousands a week stealing content from people that actually put in effort and know what they're doing.
In the last year alone I've have multiple aspiring vtubers take content I released for free and shit all over it. All because they're trying to get sponsored by Funimation/CR. So they take something we worked hard on and lie about it. Last time they took something with multiple TL notes and subtitles at the top of the screen and hid them. Then presented the subtitles at the bottom of the screen in such a way that they somehow made the offical (wrong) translation look better. In that particular scene we provided two translations. One that attempted to maintain the joke being made and another literal translation. Maybe we could have handled it better. But the one that retained the joke was kind of iffy and it was really hard to do that joke in English. So we put it at the top of the screen and a literal translation at the bottom.
In another scene they either lied or made a common mistake. Obviously they don't know Japanese. But they used it as a way to claim the CR TL was better. Since whoever did it didn't understand Japanese either.
This kind of thing isn't just limited to people trying to make youtube money either. We've had multiple people try to shut us down on places like nyaa. If you dig into who's running the place now you discover the same thing: It's all CR or former CR employees. That's why they push official translations and rips so hard these days. They're the same ones that ddos'd all the anon nyaa clones as well (after they stole the code anon provided for free to run their own).
It's a really horrible situation now on the internet. People that actually do good work are shut out of everywhere with gatekeeping. Any attempt to start up a place outside of these people's control get ddos'd into oblivion and if that fails they email your hosting provider or domain provider and shut you down that way. They steal your work, claim it as their own and then get donations from clueless people that praise them for their "hard work". These people using AI are just the same thing happening with art.
Why do people pay them and support them? No idea. I'm going to guess they're lonely and any interaction they get from a "virtual friend" brightens their day. Things are so bad we've monetized friendship. It's really sad.
No.2347
>>2333These models don't matter to me until someone trains it on booru stuff to make it palatable. There's like 50 different real life models around and I can't care enough to compare them. "Oh, that one does apples on tables better than the other one which is better at mountains."