luna-diffusion

Maintainer: proximasanfinetuning

Total Score

45

Last updated 9/6/2024

📉

PropertyValue
Run this modelRun on HuggingFace
API specView on HuggingFace
Github linkNo Github link provided
Paper linkNo paper link provided

Create account to get full access

or

If you already have an account, we'll log you in

Model overview

luna-diffusion is a fine-tuned version of Stable Diffusion 1.5 created by proximacentaurib. It was trained on a few hundred mostly hand-captioned high-resolution images to produce an ethereal, painterly aesthetic. Similar models include Dreamlike Diffusion 1.0, which is also a fine-tuned version of Stable Diffusion, and Hitokomoru Diffusion, which has been fine-tuned on Japanese artwork.

Model inputs and outputs

luna-diffusion is a text-to-image generation model that takes a text prompt as input and produces an image as output. The model was fine-tuned on high-resolution images, so it works best at 768x768, 512x768, or 768x512 pixel resolutions. The model also supports adding "painting" to the prompt to increase the painterly effect, and "illustration" to get more vector art-style images.

Inputs

  • Text prompt: A natural language description of the desired image, such as "painting of a beautiful woman with red hair, 8k, high quality"

Outputs

  • Image: A generated image matching the provided text prompt, saved as a JPEG or PNG file

Capabilities

luna-diffusion can generate high-quality, painterly-style images based on text prompts. The model produces ethereal, soft-focus images with a focus on detailed scenes and figures. It works particularly well for prompts involving people, nature, and fantasy elements.

What can I use it for?

luna-diffusion is well-suited for applications in art, design, and creative expression. You could use it to generate concept art, illustrations, or other visual assets for things like games, books, marketing materials, and more. The model's unique aesthetic could also make it useful for mood boards, visual inspiration, or other creative projects.

Things to try

To get the best results from luna-diffusion, try experimenting with different aspect ratios and resolutions. The model was trained on 768x768 images, so that size or similar ratios like 512x768 or 768x512 tend to work well. You can also play with the "painting" and "illustration" keywords in your prompts to adjust the style. Additionally, the DPM++ 2M sampler often produces crisp, clear results, while the Euler_a sampler gives a softer look.



This summary was produced with help from an AI and may contain inaccuracies - check out the links to read the original source documents!

Related Models

stable-diffusion-inpainting

runwayml

Total Score

1.5K

stable-diffusion-inpainting is a latent text-to-image diffusion model developed by runwayml that is capable of generating photo-realistic images based on text inputs, with the added capability of inpainting - filling in masked parts of images. Similar models include the stable-diffusion-2-inpainting model from Stability AI, which was resumed from the stable-diffusion-2-base model and trained for inpainting, and the stable-diffusion-xl-1.0-inpainting-0.1 model from the Diffusers team, which was trained for high-resolution inpainting. Model inputs and outputs stable-diffusion-inpainting takes in a text prompt, an image, and a mask image as inputs. The mask image indicates which parts of the original image should be inpainted. The model then generates a new image that combines the original image with the inpainted content based on the text prompt. Inputs Prompt**: A text description of the desired image Image**: The original image to be inpainted Mask Image**: A binary mask indicating which parts of the original image should be inpainted (white for inpainting, black for keeping) Outputs Generated Image**: The new image with the inpainted content Capabilities stable-diffusion-inpainting can be used to fill in missing or corrupted parts of images while maintaining the overall composition and style. For example, you could use it to add a new object to a scene, replace a person in a photo, or fix damaged areas of an image. The model is able to generate highly realistic and cohesive results, leveraging the power of the Stable Diffusion text-to-image generation capabilities. What can I use it for? stable-diffusion-inpainting could be useful for a variety of creative and practical applications, such as: Restoring old or damaged photos Removing unwanted elements from images Compositing different visual elements together Experimenting with different variations of a scene or composition Generating concept art or illustrations for games, films, or other media The model's ability to maintain the overall aesthetic and coherence of an image while manipulating specific elements makes it a powerful tool for visual creativity and production. Things to try One interesting aspect of stable-diffusion-inpainting is its ability to preserve the non-masked parts of the original image while seamlessly blending in the new content. This can be used to create surreal or fantastical compositions, such as adding a tiger to a park bench or a spaceship to a landscape. By carefully selecting the mask regions and prompt, you can explore the boundaries of what the model can achieve in terms of image manipulation and generation.

Read more

Updated Invalid Date

🤖

dreamlike-diffusion-1.0

dreamlike-art

Total Score

1.0K

Dreamlike Diffusion 1.0 is a Stable Diffusion 1.5 model fine-tuned by dreamlike.art on high-quality art. It produces dreamlike, surreal images with a distinctive visual style. The model can be used alongside other Dreamlike models like Dreamlike Photoreal 2.0 and Dreamlike Anime 1.0 to generate a variety of artistic outputs. Model inputs and outputs Dreamlike Diffusion 1.0 uses text prompts as inputs to generate unique images. The model was trained on high-quality art, so it excels at producing dreamlike, surreal visuals with a distinctive aesthetic. Outputs are high-resolution images in a variety of aspect ratios. Inputs Text prompt**: A description of the desired image, which can include details about the scene, style, and subject matter. Outputs High-resolution image**: The generated image, which can be in different aspect ratios like 2:3, 3:2, 9:16, or 16:9 for portrait or landscape orientations. Capabilities Dreamlike Diffusion 1.0 can generate a wide range of surreal, dreamlike images with a distinctive visual style. The model performs well on prompts that call for fantastical, imaginative scenes, as well as those that blend realism with more abstract or stylized elements. Users have found success combining the model with Dreamlike Photoreal 2.0 and Dreamlike Anime 1.0 to explore different artistic directions. What can I use it for? Dreamlike Diffusion 1.0 is well-suited for creative projects that require imaginative, visually striking imagery. This could include album covers, book illustrations, concept art, or promotional materials. The model's unique aesthetic makes it valuable for those looking to add a touch of the surreal or fantastical to their work. Users can also experiment with combining it with other Dreamlike models to expand their creative possibilities. Things to try Try using the dreamlikeart keyword in your prompts to further enhance the model's distinctive style. Non-square aspect ratios can also produce interesting results, so experiment with portrait and landscape orientations. Higher resolutions like 640x640px, 512x768px, or 768x512px may yield more detailed outputs as well.

Read more

Updated Invalid Date

🎯

stable-diffusion-v1-5

stable-diffusion-v1-5

Total Score

70

The stable-diffusion-v1-5 model is a latent text-to-image diffusion model capable of generating photo-realistic images from any text input. This model was fine-tuned from the Stable-Diffusion-v1-2 checkpoint with 595k additional training steps at 512x512 resolution on the "laion-aesthetics v2 5+" dataset, along with 10% dropping of the text-conditioning to improve classifier-free guidance sampling. It can be used with both the Diffusers library and the RunwayML GitHub repository. Model inputs and outputs The stable-diffusion-v1-5 model takes a text prompt as input and generates a photo-realistic image as output. The text prompt can describe any scene or object, and the model will attempt to render a corresponding visual representation. Inputs Text prompt**: A textual description of the desired image, such as "a photo of an astronaut riding a horse on mars". Outputs Generated image**: A photo-realistic image that matches the provided text prompt, in this case an image of an astronaut riding a horse on Mars. Capabilities The stable-diffusion-v1-5 model is capable of generating a wide variety of photo-realistic images from text prompts. It can create scenes with people, animals, objects, and landscapes, and can even combine these elements in complex compositions. The model has been trained on a large dataset of images and is able to capture fine details and nuances in its outputs. What can I use it for? The stable-diffusion-v1-5 model can be used for a variety of applications, such as: Art and Design**: Generate unique and visually striking images to use in art, design, or advertising projects. Education and Research**: Explore the capabilities and limitations of generative AI models, or use the model in educational tools and creative exercises. Prototyping and Visualization**: Quickly generate images to help visualize ideas or concepts during the prototyping process. Things to try One interesting thing to try with the stable-diffusion-v1-5 model is to experiment with prompts that combine multiple elements or have a more complex composition. For example, try generating an image of "a robot artist painting a portrait of a cat on the moon" and see how the model handles the various components. You can also try varying the level of detail or specificity in your prompts to see how it affects the output.

Read more

Updated Invalid Date

hitokomoru-diffusion

Linaqruf

Total Score

78

hitokomoru-diffusion is a latent diffusion model that has been trained on Japanese Artist artwork, /Hitokomoru. The current model has been fine-tuned with a learning rate of 2.0e-6 for 20000 training steps/80 Epochs on 255 images collected from Danbooru. The model is trained using NovelAI Aspect Ratio Bucketing Tool so that it can be trained at non-square resolutions. Like other anime-style Stable Diffusion models, it also supports Danbooru tags to generate images. There are 4 variations of this model available, trained for different numbers of steps ranging from 5,000 to 20,000. Similar models include the hitokomoru-diffusion-v2 model, which is a continuation of this model fine-tuned from Anything V3.0, and the cool-japan-diffusion-2-1-0 model, which is a Stable Diffusion v2 model focused on Japanese art. Model inputs and outputs Inputs Text prompt**: A text description of the desired image to generate, which can include Danbooru tags like "1girl, white hair, golden eyes, beautiful eyes, detail, flower meadow, cumulonimbus clouds, lighting, detailed sky, garden". Outputs Generated image**: An image generated based on the input text prompt. Capabilities The hitokomoru-diffusion model is able to generate high-quality anime-style artwork with a focus on Japanese artistic styles. The model is particularly skilled at rendering details like hair, eyes, and natural environments. Example images showcase the model's ability to generate a variety of characters and scenes, from portraits to full-body illustrations. What can I use it for? You can use the hitokomoru-diffusion model to generate anime-inspired artwork for a variety of purposes, such as illustrations, character designs, or concept art. The model's ability to work with Danbooru tags makes it a flexible tool for creating images based on specific visual styles or themes. Some potential use cases include: Generating artwork for visual novels, manga, or anime-inspired media Creating character designs or concept art for games or other creative projects Experimenting with different artistic styles and aesthetics within the anime genre Things to try One interesting aspect of the hitokomoru-diffusion model is its support for training at non-square resolutions using the NovelAI Aspect Ratio Bucketing Tool. This allows the model to generate images with a wider range of aspect ratios, which can be useful for creating artwork intended for specific formats or platforms. Additionally, the model's ability to work with Danbooru tags provides opportunities for experimentation and fine-tuning. You could try incorporating different tags or tag combinations to see how they influence the generated output, or explore the model's capabilities for generating more complex scenes and compositions.

Read more

Updated Invalid Date