Xorte logo

News Markets Groups

USA | Europe | Asia | World| Stocks | Commodities



Add a new RSS channel

 
 


Keywords

2024-10-22 20:40:22| Engadget

Stable Diffusion, an open-source alternative to AI image generators like Midjourney and DALL-E, has been updated to version 3.5. The new model tries to right some of the wrongs (which may be an understatement) of the widely panned Stable Diffusion 3 Medium. Stability AI says the 3.5 model adheres to prompts better than other image generators and competes with much larger models in output quality. In addition, its tuned for a greater diversity of styles, skin tones and features without needing to be prompted to do so explicitly. The new model comes in three flavors. Stable Diffusion 3.5 Large is the most powerful of the trio, with the highest quality of the bunch, while leading the industry in prompt adherence. Stability AI says the model is suitable for professional uses at 1 MP resolution. Meanwhile, Stable Diffusion 3.5 Large Turbo is a distilled version of the larger model, focusing more on efficiency than maximum quality. Stability AI says the Turbo variant still produces high-quality images with exceptional prompt adherence in four steps. Finally, Stable Diffusion 3.5 Medium (2.5 billion parameters) is designed to run on consumer hardware, balancing quality with simplicity. With its greater ease of customization, the model can generate images between 0.25 and 2 megapixel resolution. However, unlike the first two models, which are available now, Stable Diffusion 3.5 Medium doesnt arrive until October 29. The new trio follows the botched Stable Diffusion 3 Medium in June. The company admitted that the release didnt fully meet our standards or our communities expectations, as it produced some laughably grotesque body horror in response to prompts that asked for no such thing. Stability AIs repeated mentions of exceptional prompt adherence in todays announcement are likely no coincidence. Although Stability AI only briefly mentioned it in its announcement blog post, the 3.5 series has new filters to better reflect human diversity. The company describes the new models human outputs as representative of the world, not just one type of person, with different skin tones and features, without the need for extensive prompting. Lets hope its sophisticated enough to account for subtleties and historical sensitivities, unlike Googles debacle from earlier this year. Unprompted to do so, Gemini produced collections of egregiously inaccurate historical photos, like ethnically diverse Nazis and US Founding Fathers. The backlash was so intense that Google didnt reincorporate human generations until six months later.This article originally appeared on Engadget at https://www.engadget.com/ai/stable-diffusion-35-follows-your-prompts-more-closely-and-generates-more-diverse-people-184022965.html?src=rss


Category: Marketing and Advertising

 

Latest from this category

18.02Samsung teases mobile AI photography tools ahead of Unpacked
17.02Google I/O 2026 is set for May 19 and 20
17.02Texas AG sues TP-Link over purported connection to China
17.02Netflix is adapting the board game Ticket to Ride
17.02WordPress adds an AI assistant
17.02Netflix is streaming its first MMA fight on May 16
17.02The first full trailer for The Mandalorian and Grogu is here
17.02Kingdom Come: Deliverance 2 and The Witcher 3 are coming to Game Pass
Marketing and Advertising »

All news

18.02Illegal skin lightening cream being sold in UK butchers, watchdog warns
18.02Samsung teases mobile AI photography tools ahead of Unpacked
17.02I invested 12,000 in Brewdog - I think I've lost it all
17.02Tomorrow's Earnings/Economic Releases of Note; Market Movers
17.02Bull Radar
17.02Bear Radar
17.02Stocks Reversing Slightly Higher into Final Hour on Stable Long-Term Rates, Earnings Outlook Optimism, Technical Buying, Transport/Biotech Sector Strength
17.02What Makes This Trade Great? CMPS
More »
Privacy policy . Copyright . Contact form .