{ "_id": "621ffdc136468d709f17ea61", "id": "openai/clip-vit-base-patch16", "private": false, "pipeline_tag": "zero-shot-image-classification", "library_name": "transformers", "tags": [ "transformers", "pytorch", "jax", "clip", "zero-shot-image-classification", "vision", "arxiv:2103.00020", "arxiv:1908.04913", "endpoints_compatible", "region:us" ], "downloads": 4931979, "likes": 115, "modelId": "openai/clip-vit-base-patch16", "author": "openai", "sha": "57c216476eefef5ab752ec549e440a49ae4ae5f3", "lastModified": "2022-10-04T09:42:28.000Z", "gated": false, "disabled": false, "widgetData": [ { "src": "https://huggingface.co/datasets/mishig/sample_images/resolve/main/cat-dog-music.png", "candidate_labels": "playing music, playing sports", "example_title": "Cat & Dog" } ], "model-index": null, "config": { "architectures": [ "CLIPModel" ], "model_type": "clip", "tokenizer_config": { "unk_token": { "content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken" }, "bos_token": { "content": "<|startoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken" }, "eos_token": { "content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken" }, "pad_token": "<|endoftext|>" } }, "cardData": { "tags": [ "vision" ], "widget": [ { "src": "https://huggingface.co/datasets/mishig/sample_images/resolve/main/cat-dog-music.png", "candidate_labels": "playing music, playing sports", "example_title": "Cat & Dog" } ] }, "transformersInfo": { "auto_model": "AutoModelForZeroShotImageClassification", "pipeline_tag": "zero-shot-image-classification", "processor": "AutoProcessor" }, "siblings": [ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "flax_model.msgpack" }, { "rfilename": "merges.txt" }, { "rfilename": "preprocessor_config.json" }, { "rfilename": "pytorch_model.bin" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "vocab.json" } ], "spaces": [ "merve/EVACLIP", "vishnun/CLIPnCROP", "taesiri/CLIPScore", "merve/compare_clip_siglip", "sayakpaul/evaluate-sd-schedulers", "navervision/LinCIR", "Baron-GG/LLAUS", "sayakpaul/demo-custom-css", "Thuchk/Stable_diffusion_v1.4", "oscurantismo/diane", "aim9061/sentiment-analysis", "hafidhsoekma/Hunter-X-Hunter-Anime-Classification", "AbdullahMd12/openai-clip-vit-base-patch16", "Alivin/openai-clip-vit-base-patch16", "jke94/demo-openai-clip-vit", "kushinm/human_vs_AI_caption_challenge", "bczhou/clip-gpt2", "RidaFatima/RIDAAIPROJECT", "Tharunika1601/text2speech", "ToletiSri/Capstone", "venkat-natchi/AnyModeAssistant", "Vasudevakrishna/CapstoneERA", "HemaAM/Multimodal_GPT", "Vasudevakrishna/TAI2T_Multimodel", "Vasudevakrishna/TSAI_Capstone", "drxlouis/opdracht-ai-interface", "VarunSivamani/CLIP", "dincali/text-image-converter", "ConceptaMAGIC/operations-CLIP-Score", "Shuaizhang7/Surveillance_scene_recognition", "sominyeo0131/gkstkfkdduthalsru", "drxlouis/AI_Interface", "Xuroo/openai-clip-vit-base-patch16", "Vasudevakrishna/MultiModel_LLM_ERAV2", "swithana/openai-clip-vit-base-patch16", "qbenqa/recomp", "Nikhil123098/openai-clip-vit-base-patch16", "Shahabmoin/Real-Time-Image-to-Text-Generator", "UPstud/ColorCtrl", "Ahmed1871992/Personalized-ImageGen", "gskdsrikrishna/Image_Gen", "joy1515/retrieval-ai", "siro1/triton-for-llms", "tennant/old_tok" ], "createdAt": "2022-03-02T23:29:05.000Z", "inference": "warm", "usedStorage": 2992733617 }