Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Text-to-Image Alignment Performance of the pixart-sigma Model #25

Open
xiexiaoshinick opened this issue Apr 10, 2024 · 4 comments
Open
Labels
good first issue Good for newcomers

Comments

@xiexiaoshinick
Copy link

First of all, I would like to express my gratitude for your open-source pixart-sigma project. As a developer who has been closely following your work, I couldn't wait to test the new model as soon as it was released. I used the GenEval framework to evaluate the model's performance in text-to-image alignment. The results showed that compared to SDXL and PlayGroundv2.5, there is still room for improvement in this aspect.

模型 Overall single two counting colors position color_attr
SD1.5 42.34 95.62 37.63 37.81 74.73 3.50 4.75
SD1.5-DPO 43.00 96.88 39.90 38.75 75.53 3.25 3.75
SDXL 55.63 98.12 75.25 43.75 89.63 11.25 15.75
playgroundv2.5 56.37 97.81 77.02 51.88 83.78 11.00 16.75
SDXL-DPO 58.02 99.38 82.58 49.06 85.11 13.50 18.50
PixArt-⍺(1024) 47.16 97.81 46.21 45.00 77.93 9.00 7.00
PixArt-Σ (512) 52.03 98.12 59.02 50.62 80.05 9.75 15.50
PixArt-Σ (1024) 54.39 98.44 62.88 49.69 82.45 12.00 20.00

I noticed that Stable Diffusion 3 adopted the DPO (Direct Preference Optimization) method, which greatly improved the text-to-image alignment. In this regard, I would like to ask if your team has any plans to incorporate similar optimization methods in future versions to further enhance the model's performance in this area.

@lawrence-cj
Copy link
Contributor

Thank you so much for your work and help. DPO will definitely help to get consistent improvement. Actually, we would prefer to encourage our community members to do their specific DPO, not just do everything on our own~.

@lawrence-cj lawrence-cj added the good first issue Good for newcomers label Apr 10, 2024
@ApolloRay
Copy link

Thank you so much for your work and help. DPO will definitely help to get consistent improvement. Actually, we would prefer to encourage our community members to do their specific DPO, not just do everything on our own~.

I will try this. However, the sigma-DMD will be released ?

@lawrence-cj
Copy link
Contributor

Already released! Refer to the readme:)

@ApolloRay
Copy link

First of all, I would like to express my gratitude for your open-source pixart-sigma project. As a developer who has been closely following your work, I couldn't wait to test the new model as soon as it was released. I used the GenEval framework to evaluate the model's performance in text-to-image alignment. The results showed that compared to SDXL and PlayGroundv2.5, there is still room for improvement in this aspect.

模型 Overall single two counting colors position color_attr
SD1.5 42.34 95.62 37.63 37.81 74.73 3.50 4.75
SD1.5-DPO 43.00 96.88 39.90 38.75 75.53 3.25 3.75
SDXL 55.63 98.12 75.25 43.75 89.63 11.25 15.75
playgroundv2.5 56.37 97.81 77.02 51.88 83.78 11.00 16.75
SDXL-DPO 58.02 99.38 82.58 49.06 85.11 13.50 18.50
PixArt-⍺(1024) 47.16 97.81 46.21 45.00 77.93 9.00 7.00
PixArt-Σ (512) 52.03 98.12 59.02 50.62 80.05 9.75 15.50
PixArt-Σ (1024) 54.39 98.44 62.88 49.69 82.45 12.00 20.00
I noticed that Stable Diffusion 3 adopted the DPO (Direct Preference Optimization) method, which greatly improved the text-to-image alignment. In this regard, I would like to ask if your team has any plans to incorporate similar optimization methods in future versions to further enhance the model's performance in this area.

I meet this problem, durring use the GenEval,
ordering = np.argsort(bbox[index][:, 4])[::-1] TypeError: 'DetDataSample' object is not subscriptable
Have you met this problem ?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
good first issue Good for newcomers
Projects
None yet
Development

No branches or pull requests

3 participants