cover of episode Unleashing Firefly Image 3: Adobe's latest AI innovation

Unleashing Firefly Image 3: Adobe's latest AI innovation

2024/4/24
logo of podcast No Priors AI

No Priors AI

AI Deep Dive AI Chapters Transcript
People
主持人
专注于电动车和能源领域的播客主持人和内容创作者。
Topics
Adobe发布了新的AI图像生成模型Firefly Image 3,该模型具有许多令人印象深刻的功能,例如生成逼真的图像,其质量虽然不如Midjourney,但在易用性方面具有优势。Firefly Image 3专注于加快创意过程,用户可以通过清晰的菜单选项快速生成图像,而无需学习复杂的参数。它具有更好的照片真实感,可以从单个提示生成各种风格的输出(照片、艺术、插图等),并可以设置情绪或灯光。此外,Firefly Image 3还具有参考图像功能,允许用户上传图像并基于提示生成类似的图像,简化了提示编写过程。该功能类似于ChatGPT根据示例文本生成文本的方式。虽然该功能要求用户拥有上传图像的权利,但Adobe正在努力阻止未经授权的图像被用作参考,并声称该功能安全用于商业用途。Firefly Image 3已集成到Photoshop中,并包含其他生成式AI工具,例如生成背景和生成类似图像的功能,其图像生成质量比以前版本更高,并包含一个新的调整画笔工具,类似于Lightroom中的工具。Adobe的优势在于其庞大的用户基础和已有的软件工具,这使得Firefly Image 3能够迅速被广泛使用。

Deep Dive

Chapters
This chapter introduces Adobe's new AI image generator, Firefly Image 3, highlighting its ability to create photorealistic images and comparing it to Midjourney. It discusses the user-friendly interface and features like aspect ratio adjustment and content type selection.
  • Firefly Image 3 excels at generating realistic images.
  • It offers a user-friendly interface with clear menu options.
  • It allows for aspect ratio and content type adjustments.

Shownotes Transcript

Translations:
中文

Adobe has just announced a brand new AI model and that is a new image generator. It is the Firefly Image 3. This thing has a bunch of amazing features. I'm going to be going over all of them, showing you some demos. I actually got my hands on using this product recently and I've been really impressed with a lot of things that Adobe has done. So they did a big press release that kind of outlined everything going on here.

But, you know, the funny thing is where I get a lot of my news these days is literally just from X. So I actually saw this first on X. If you're interested in following like an influencer that is in Adobe, Chris Castuanova is an awesome influencer. She works at Adobe and she always is kind of tweeting about the latest things. Anyways, she shared a really interesting video recently of...

some images that this thing can do. And the main thing she said is that it's good at generating realistic images. So this kind of new version, and we saw a very similar kind of progress with like Midjourney where it started being really good at kind of fantasy. And then it's kind of good at kind of like

different types, styles of art. And eventually this thing is able to just straight up generate what looks very similar to like realistic photos, photo realistic photos. This is, I think one of the big things that they are focusing on. So people have been commenting on this. Um, different people have been like doing their own tests and showing, you know, what they were able to generate with it. Uh, people have, you know, examples of, uh,

Models with blue hair and blue dresses and whatever but like they look photorealistic like an actual person. So it's It's very very good Someone asked an interesting question and I'm gonna cover this because I didn't see this in any press releases or anywhere else So this is it feels like Twitter is kind of the only place you get this someone said do you think it has character consistency? Chris said

There's style and structure consistency. There's a way you can trick it to some sort of character consistency. She has like a tutorial and video where she kind of showed you how to do that. That being said, this is not something that specifically does. So definitely Midjourney is still ahead of the curve on this because Midjourney recently unveiled a feature where you have character consistency, meaning that you could have the same character

in different images and different places. So that is definitely something that Midjourney has ahead, which is kind of a big deal if you're trying to do, you know, some sort of like book or some sort of graphics on your website where you need like the same character in multiple different positions or places or shots.

This is something that Midjourney is now able to do, does it fairly well. I think it kind of made this big announcement of the ability to do this back in March. So Midjourney has been doing this for a minute. It seems like Adobe is sort of catching up on a lot of regards, but it doesn't seem like this one is something they've specifically caught up with quite yet. So hopefully this is something they're going to be able to do. One thing that I wanted to say is there was an interesting blog about this whole thing on Tom's Guide.

And Ryan Morrison published an article on this. Just one thing I thought was interesting. He says, "Adobe says the work on Firefly 3 was focused on speeding up ideation, allowing designers to go from an idea to a fully fledged image in as little time and with as little friction as possible. I think they've achieved it. Unlike Midjourney where you have to learn multiple parameters and how to implement them, Firefly has a series of well-defined and clear menu options." Now, this is something true. I've tested this out and I'll show you some demos.

But they do have a lot more menu options, right? So think drop downs where it's like, what style do you want? Is it ultra realistic, right? So they kind of have those, which feels a little bit,

you know, it feels like there's less learning how to be less hardcore, really good prompt engineer. One other thing he said was Firefly 3 seems to have better photo realism, a wider variety of outputs from a single prompt across styles like photo, art, illustration, as well as options to set mood or lighting. So it's got a lot of really cool features. When I was playing with it,

I actually asked it to generate something. I've done a couple things. The first one I asked it to do was to generate a picture of like a house and a cottage. It's kind of this default prompt and it did a great job. The one interesting thing I will say is I put the words hyper-realistic, which if you do that on...

If you do that on mid journey, it was hyper realistic. The image originally generated for me was not. And then again, I kind of changed it and asked for like a castle. I said a wooden castle and said for it to be hyper realistic, said for there to be statues. It actually generated, it looks like a photograph. Is it perfect? I would say probably not, but like it actually looks pretty dang good. There's a person in the background, there's statues, there's like mountains, there's a

there's a castle with like a bunch of wooden elements to the castle, which is what I asked for, which so I thought it was kind of cool. And of course it does, you know, like a lot of these image generators, it generated four variations. And I will say, you know, some of them were better than others. Some of them, the windows were like filled in with bricks and others looked a little bit more realistic. So yeah, you kind of get what you pay for. But I think as far as

As far as this goes, similar to mid journey, you're going to be able to generate a couple. Now, one other thing that I will say, you have the ability to change the aspect ratio if you want it to generate squares or if you want it to generate like portrait, widescreen, landscape. That's really nice. I don't see that on all of them. Content type, you can get it to do a photo or more an artistic style and you can kind of toggle that on and off so you can see what you want it to do.

So, yeah, there's a bunch of really interesting things. I wanted to go over some of the very specific features that they've actually unveiled.

One of the big ones is called reference image. So with reference image, you're actually able to upload an image and get it to generate something similar based off of your prompt. So here's a quote from Ellie Greenfield, who's a chief technology officer for digital media over at Adobe. And Ellie said, prompting is a pain in the butt. Why spend an hour trying to craft a three paragraph prompt? If you have an image that you've created that's a,

that's exactly the thing you wanna reference. That saying a picture's worth a thousand words applies here, right? So you're actually gonna be able to be like,

I have this great picture of a castle. Now I want it to generate me a picture of a castle, but to change this, the element or that element, you can literally use your own images to instead of a prompt, which I think is amazing because I do this exact same thing with chat GPT. For example, it's like I want it to write me a specific type of tweet or a specific paragraph of my article. I don't want to think of like how to like what my style of writing is to get it to help me clone that. I just want to say like,

write it like me. So what I always do is I just take an entire, you know, article and plug it into my prompt. And I'm like, Hey, write me a paragraph about X, Y, and Z. Do it in the style of, in my previous, you know, writing style. Here's a excerpt that I've previously written. It just reads that. And then boom, it'll output something that's similar to how I've written in the past. Um,

This is amazing. It wasn't really possible to do this with images, so Adobe is making it possible. Now, there's a lot of people that are kind of concerned about it for a couple different reasons. So I want to read a paragraph from an article by The Verge that kind of discusses this. They say, users are expected to have the rights to use images they want to reference. Greenfield told The Verge that a message will flag this...

that a message will flag this ownership requirement when the tool is first used and that the company is working on a universal do not train tag for Adobe's content authenticity initiative that it's also going to block images from being used as reference. So images uploaded as reference materials won't be used to train Firefly. Okay, first off, I think that's great. You know, it would be super annoying if you're like uploading a picture and take

to do something and then like everything upload is just getting sucked into their giant data set. I don't think people want that, especially corporations. So that is, I think, a good thing. They also said, quote, despite the ownership responsibility being placed on users, Adobe says this new referencing tool is still, quote, safe for commercial use. One of the most notable advantages that Adobe claims Firefly has over rival companies.

generative AI models. So they're saying this is still good for commercial use. Um, everything is trained with their dataset. Adobe has been really big on this. And so I think this is going to be interesting. All of this is getting baked into Photoshop, um, which is again, really exciting. I, there have a couple other generative AI tools that are kind of, uh,

already in here and a lot of the stuff is stuff we've been seeing, right? They have like a generative AI generate background where you can like upload a picture of your product, for example, and then it can come up with a bunch of different backgrounds. You can say like, you know, they have an example where it's like a perfume bottle and then they're saying like generate a background with pink peonies or orchids or water and it's like changing the background behind this perfume bottle. They have another one called generate similar. This is essentially going to be used, um,

where it can generate images similar or come up with different variations of the same image, which is really interesting. And then their third, this new model they have has a lot higher quality of capabilities in generating these images compared to all of the, you know, its previous ones. So it's going to be able to do a lot of interesting things. They have something called an adjustment brush that they've recently added. They showed a demo where like

Essentially, honestly, it reminds me of something that's already in kind of like Adobe Shucks, I forget. Whatever the one all the photographers use where they're creating the filters and stuff. Lightroom, there you go. And essentially, they have like a brush where they highlight someone's glasses and they can change the color of his glasses with it. So, I mean, that feels kind of like an Adobe tool that you've seen other places. But this is something that's also gonna be baked into this.

All in all, I think this is really impressive. I mean, the big news here is just how much better this model is getting at generating realistic like photos, not perfect. And I don't actually think it's as good as Midjourney right now. It's kind of interesting when you hear people talk about this, they're saying like, this is so much better than Midjourney because you don't have to learn these complex prompts. Like, yeah, but...

But at the same time, the quality of Midjourney, I think is still a little bit better. So like maybe it's a little bit more complex to learn the prompts, but if you want like the best in class, it's gonna be Midjourney. So all in all, I think that this is a really interesting time to be making these photos. I'll keep you up to date as they roll this out

and add new features. But all in all, I think this is an amazing update, great work from Adobe really keeping up with this new image model here. And the thing that's amazing about, of course, Photoshop or Adobe that you're not going to see with other people even like Midjourney is they already have such popular software tools in the creative space and image and video generation, or editing tools that now that they can plug this in, that's

millions and millions of users that are going to get direct access to this. So I do think, you know, you can't discount Adobe. You can't count them out because they have such a huge user base. This is already plugged in and they just, you know, integrate this into a tool and millions of people will instantly be using it. If you enjoyed the episode today, I would really, really appreciate it if you could drop a review, like the video if you're watching on YouTube or follow us on Spotify and Apple, leave us a review. I love them all. Hope that you all have an amazing rest of your day.