Skip to content

Preprocessing of pretraining data #43

@seanzhuh

Description

@seanzhuh

Hi, Thanks for open-sourcing such a wonderful work.

Regarding the preprocessing of pretraining data, did you apply this template prompts in ULIP to both the raw texts, blip and msft generated captions, but not the retrieved texts right? As I can tell, in pretraining data, each *.npy file contains embeddings of "original" and "prompt_avg" versions for text_feat, blip_caption_feat, msft_caption_feat, only the retrieval_text_feat does not have "prompt_avg" version.

And could you please give me a hint on where to download the thumbnail images you used to extract the thumbnail feats as the thumbnail images are not included in released pretraining data, only the extracted thumbnail embeddings are available.

If you can provide your full preprocess file to extract text and image embeddings, it would be of great help!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions