FashionGAN: Difference between revisions

From FDHwiki
Jump to navigation Jump to search
No edit summary
Line 120: Line 120:


StyleGAN is a type of generative adversarial network (GAN) developed by NVIDIA that's primarily used for generating high-quality, realistic images. It's renowned for its ability to create lifelike human faces, animals and objects. On all versions of StyleGAN that exists, we chose to work with StyleGAN2-ADA (https://github.com/NVlabs/stylegan2-ada-pytorch), based on StyleGAN2. It is less powerfull than StyleGAN3, but regarding our ressources a model based on StyleGAN2 is more appropriate. ADA stands for "adaptive discriminator augmentation", it signifies that this model is better suited for smaller datasets (for example, StyleGAN2-ADA aims to have the same results than the original model with a 30k datasets instead of 100k).
StyleGAN is a type of generative adversarial network (GAN) developed by NVIDIA that's primarily used for generating high-quality, realistic images. It's renowned for its ability to create lifelike human faces, animals and objects. On all versions of StyleGAN that exists, we chose to work with StyleGAN2-ADA (https://github.com/NVlabs/stylegan2-ada-pytorch), based on StyleGAN2. It is less powerfull than StyleGAN3, but regarding our ressources a model based on StyleGAN2 is more appropriate. ADA stands for "adaptive discriminator augmentation", it signifies that this model is better suited for smaller datasets (for example, StyleGAN2-ADA aims to have the same results than the original model with a 30k datasets instead of 100k).
= Copyright =
Copyright question


= Github =  
= Github =  


[https://github.com/shayayan/FDH-FashionGAN/tree/main FDH-FashionGAN]
[https://github.com/shayayan/FDH-FashionGAN/tree/main FDH-FashionGAN]

Revision as of 13:52, 17 December 2023

Introduction

Je me suis mis en groupe avec Romane, erreur.

Motivation

The impetus behind embarking on this groundbreaking project stemmed from a collective desire to revolutionize the creative process for fashion designers. Recognizing the ever-evolving landscape of the fashion industry, our team felt compelled to innovate and address a common challenge: the need for novel inspiration and streamlined design techniques.

The foundation of our motivation lies in leveraging the vast potential of artificial intelligence alongside an extensive repository of fashion show data. The aim? To empower designers with a cutting-edge tool that transcends conventional boundaries. By harnessing AI capabilities, we envisioned a tool that not only generates fresh and inventive clothing designs but also acts as a wellspring of inspiration for designers seeking a new creative direction.

The core essence of our project centers on the creation of an AI-driven platform capable of producing unique clothing visuals. Through the amalgamation of innovative technologies like DragGAN and AutoEncoder, designers will not only access generated designs but also possess the ability to customize and refine these creations to match their artistic vision.

Ultimately, our aspiration is to provide fashion designers with an invaluable resource that nurtures creativity, encourages experimentation, and accelerates the design process. By enabling designers to explore limitless possibilities, our tool aspires to be a catalyst for transformative innovation within the fashion sphere.

Deliverables

Dataset

A folder with ((insert number)) of images from ((insert number)) of fashion shows, scraped from [www.nowfashion.com NowFashion].

Software

Milestones and Project plan

Milestones

Milestone 1: DragGAN

  • Understand how DragGAN works
  • Find a dataset that would be appropriate for our utilization
  • Train StyleGAN

Milestone 2: Texture swap

  • Find a way to apply a texture change on an image
  • Train the Swapping Autoencoder for Deep Image Manipulation
  • Implement the Texture swap interface in our project


Milestone 3: User Interface

  • Change DragGAN's interface to make it more intuitive for our project

Milestone 4: Deliverables

  • Deliver the code on Github
  • Write the wiki page
  • Prepare the presentations

Project Plan

Week Tasks Completion
Week 3
  • Choose our project between the ones that were presented in class
  • Prepare for the first presentation
Weeks 4-5-6
  • Find some clothing datasets
  • Read papers, understand how DragGAN works
  • Define the project
  • Get familiar with the libraries we want to use (selenium, opencv...)
Weeks 7-8-9
  • Scrap the data from nowfashion website
  • Resize images
  • Clean dataset (remove the pictures that aren't in the same format as the other)
  • Get familiar with StyleGAN and its functioning
  • Find a textile dataset
Week 10
  • Define a precise planning for the following weeks
  • Prepare for the presentation
  • Write the Project Plan and the Milestones on the wiki page
  • Figure out how to implement the texture changes
  • Begin the training of the dataset with StyleGAN
Week 11
  • Install a Virtual Machine to use linux
  • Finish the training on the VM
  • Get DragGAN to work on the VM
Week 12
  • Work on the texture changing tool
  • Implement the texture changing interface
  • Start writing the description of the methods on the wiki page
Week 13
  • Adapt the DragGAN GUI to be more intuitive
  • Finalize and clean the code to then commit it to Github
  • Finish writing the wiki page

Methods

For our project, we're basing ourselves on the DragGAN project (https://github.com/XingangPan/DragGAN), into which we're importing a StyleGAN model that has trained our database.

Scraping

DragGAN

DragGAN is a pioneering deep learning model designed for unprecedented controllability in image synthesis. Unlike prior methods reliant on 3D models or supervised learning, DragGAN empowers users to interactively manipulate images by clicking handle and target points, moving them precisely within the image. By utilizing GAN feature spaces, this approach allows for diverse and precise spatial attribute adjustments across various object categories. The model enables efficient, real-time editing without additional networks, facilitating interactive sessions for layout iterations. Evaluated extensively across diverse datasets, DragGAN showcases superior manipulation effects, deforming images respecting underlying object structures, outperforming existing methods in both point tracking and image editing.

StyleGAN

StyleGAN is a type of generative adversarial network (GAN) developed by NVIDIA that's primarily used for generating high-quality, realistic images. It's renowned for its ability to create lifelike human faces, animals and objects. On all versions of StyleGAN that exists, we chose to work with StyleGAN2-ADA (https://github.com/NVlabs/stylegan2-ada-pytorch), based on StyleGAN2. It is less powerfull than StyleGAN3, but regarding our ressources a model based on StyleGAN2 is more appropriate. ADA stands for "adaptive discriminator augmentation", it signifies that this model is better suited for smaller datasets (for example, StyleGAN2-ADA aims to have the same results than the original model with a 30k datasets instead of 100k).

Copyright

Copyright question

Github

FDH-FashionGAN