# wise **Repository Path**: daoos_admin/wise ## Basic Information - **Project Name**: wise - **Description**: 图片、视频、音频局部搜索 - **Primary Language**: Python - **License**: Apache-2.0 - **Default Branch**: wise2 - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2025-12-12 - **Last Updated**: 2025-12-12 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README

WISE 2 - WISE Search Engine

wise-logo
WISE is a search engine for images, videos, and audio powered by multimodal AI, allowing you to quickly and easily search through large collections of audiovisual media. You can search using natural language, an uploaded image/audio file, or a combination of these modalities. Use WISE locally on your own collections of images/videos.

For more details, visit https://www.robots.ox.ac.uk/~vgg/software/wise/

## Key Features
Natural language search

Use natural language to describe what you want to search for.

WISE uses a language model to understand the meaning behind your query, allowing you to flexibly describe what you are looking for. Moreover, WISE uses a vision model to understand what's being depicted in an image/video (i.e. it searches on visual content rather than metadata such as keywords, tags, or descriptions), so the images do not need to be manually tagged or labelled with text captions.

Visual similarity search

Upload an image or paste an image link to find similar images:

Multi-modal search

Combine images and text in your query. For example, if you upload a picture of a golden retriever and enter the text "in snow", WISE will find images of golden retrievers in snow.

Various multimodal / vision-language models supported

Various models are supported including vision-language models from OpenCLIP (including OpenAI CLIP) and the Microsoft CLAP audio-language model.

Different ways to perform searches

Searches can be performed via:

(Note: currently the search functionality in the CLI may be missing some features.)

Safety features
## Roadmap We are planning on implementing the following features soon. Stay tuned! ## Colab demo Try out WISE in Google Colab below (Google Colab is a free cloud service allowing you to run machine learning models and code without installing anything locally): Open In Colab ## Documentation The WISE open source software is developed and maintained by the Visual Geometry Group ([VGG](https://www.robots.ox.ac.uk/~vgg/software/wise/)) at the University of Oxford. Here are some documents for users and developers of WISE. - [Install](docs/Install.md) : describes the process for installing WISE - [User Guide](docs/UserGuide.md) : demonstrates the usage of WISE using a sample video dataset - [Metadata](docs/Metadata.md) : describes support for text metadata search in WISE - Evaluation - [Multi-Instance Video Retrieval](docs/Retrieval-Evaluation.md) - Developer Resources - [Data Loading](docs/data-loading.md): describes interface for loading media files - [Feature Extractor](docs/FeatureExtractor.md) : guide for creating new feature extractors in WISE - [FeatureStore](docs/FeatureStore.md) : describes the data structure containing the extracted features - [Frontend](frontend/README.md) : describes the frontend web-based interface - [Database](src/db/README.md) : describes the structure of the internal metadata database, which stores information about the source collections (i.e. input folders), media files (e.g. images, videos, or audio files), vectors, and extra metadata - [Tests](docs/Tests.md) : describes the software testing process for WISE ## Contact Please submit any bug reports and feature requests on the [Issues page](https://gitlab.com/vgg/wise/wise/-/issues). For any queries or feedback related to the WISE software, contact [Prasanna Sridhar](mailto:prasanna@robots.ox.ac.uk), [Horace Lee](mailto:horacelee@robots.ox.ac.uk) or [Abhishek Dutta](mailto:adutta@robots.ox.ac.uk). ## Acknowledgements Development and maintenance of WISE software has been supported by the following grant: Visual AI: An Open World Interpretable Visual Transformer (UKRI Grant [EP/T028572/1](https://gow.epsrc.ukri.org/NGBOViewGrant.aspx?GrantRef=EP/T028572/1))