Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

When do you plan to release images and VR? #1

Closed
Franck-Dernoncourt opened this issue Jun 23, 2020 · 5 comments
Closed

When do you plan to release images and VR? #1

Franck-Dernoncourt opened this issue Jun 23, 2020 · 5 comments

Comments

@Franck-Dernoncourt
Copy link

No description provided.

@seo-95
Copy link
Contributor

seo-95 commented Jun 24, 2020

Hi @Franck-Dernoncourt ,
I have the same doubt before reading the paper. Real images and VR are not part of the dataset, instead they are replaced by logs annotations inside the conversations. Each log contains the ID of the furniture/clothing contained in the VR/image, in such a way the computer vision side of the challenge is totally skipped. To better understand I suggest you to read section 3.3 of the paper

@Franck-Dernoncourt
Copy link
Author

Thanks @seo-95 . Just to make sure I understand: we're given some image/VR "catalog description and other attributes" (quoting section 3.3), but not the images/VR themselves?

https://github.com/facebookresearch/simmc/tree/master/data/simmc_fashion/*.json contains image_id so I thought there could be some way to retrieve the images. Also, https://arxiv.org/pdf/2006.01460.pdf's figure 4 seems to take some raw image as input ($M_t$). But reading https://arxiv.org/pdf/2006.01460.pdf's section 8 (SIMMC-Furniture (VR) and SIMMC-Fashion (Image) paragraphs) seems to confirm your previous message:

We extract the representations for each fashion item using concatenated GloVe embeddings of its attributes (similar to SIMMC-Furniture) in addition to learning the source embedding (‘memory’ or ‘current’ item), as the multimodal context $M_t$

@seo-95
Copy link
Contributor

seo-95 commented Jun 25, 2020

Thanks @seo-95 . Just to make sure I understand: we're given some image/VR "catalog description and other attributes" (quoting section 3.3), but not the images/VR themselves?

https://github.com/facebookresearch/simmc/tree/master/data/simmc_fashion/*.json contains image_id so I thought there could be some way to retrieve the images. Also, https://arxiv.org/pdf/2006.01460.pdf's figure 4 seems to take some raw image as input ($M_t$). But reading https://arxiv.org/pdf/2006.01460.pdf's section 8 (SIMMC-Furniture (VR) and SIMMC-Fashion (Image) paragraphs) seems to confirm your previous message:

We extract the representations for each fashion item using concatenated GloVe embeddings of its attributes (similar to SIMMC-Furniture) in addition to learning the source embedding (‘memory’ or ‘current’ item), as the multimodal context $M_t$

I have just seen the image_id you mentioned and it is really strange. Maybe it's there for future possible expansion of the dataset (the introduction of images/VR would be great). I hope for an answer from Facebook organizers

@Franck-Dernoncourt
Copy link
Author

Franck-Dernoncourt commented Jun 25, 2020

@shubhamagarwal92 pointed me to https://github.com/facebookresearch/simmc/tree/master/data#overview-of-the-dataset-repository (mirror), which mentions:

images: ./simmc-furniture/figures/{object_id}.png

I agree, at that point we need the organizers to clarify if images/VRs will be released, and if so, when.

@Franck-Dernoncourt
Copy link
Author

According to an email from the organizers, no plan to release images, and see the metadata csv for the VR.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants