blip-2-vision-language
Vision-language pre-training framework bridging frozen image encoders and LLMs. Use when you need image captioning, visual question answering, image-text retrieval, or multimodal chat with state-of-the-art zero-shot performance.
Packaged view
This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.
Install command
npx @skill-hub/cli install davila7-claude-code-templates-multimodal-blip-2
Repository
Skill path: cli-tool/components/skills/ai-research/multimodal-blip-2
Vision-language pre-training framework bridging frozen image encoders and LLMs. Use when you need image captioning, visual question answering, image-text retrieval, or multimodal chat with state-of-the-art zero-shot performance.
Open repositoryBest for
Primary workflow: Design Product.
Technical facets: Full Stack, Designer.
Target audience: Development teams looking for install-ready agent workflows..
License: Unknown.
Original source
Catalog source: SkillHub Club.
Repository owner: davila7.
This is still a mirrored public skill entry. Review the repository before installing into production workflows.
What it helps with
- Install blip-2-vision-language into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
- Review https://github.com/davila7/claude-code-templates before adding blip-2-vision-language to shared team environments
- Use blip-2-vision-language for development workflows
Works across
Favorites: 0.
Sub-skills: 0.
Aggregator: No.