Pytorch implementation of 'Pratical Sampling-based Bayesian Inference for multimodal distribution'
-
Updated
Jan 29, 2019 - Jupyter Notebook
Pytorch implementation of 'Pratical Sampling-based Bayesian Inference for multimodal distribution'
Collaborative generation of unique audiovisual experiences using NFC identity cards
Todo o conteúdo produzido para a unidade curricular PF (Projeto FEUP), para o curso em Engenharia Informática e Computação na FEUP
Multitasking multimodal AI material that focus on human interaction and assistance
Public repo for the paper: "COSMic: A Coherence-Aware Generation Metric for Image Descriptions" by Mert İnan, Piyush Sharma, Baber Khalid, Radu Soricut, Matthew Stone, Malihe Alikhani
Code for IEEE MultiMedia Paper "Modeling Incongruity between Modalities for Multimodal Sarcasm Detection."
Interpolate between two text concepts using a CLIP model and FiftyOne Plugins!
Distributed computing framework for Multimodal data written in Python
SCOTCH is a Single-Cell multi-modal integration method leveraging the Optimal Transport algorithm and a cell matCHing strategy
Engage in a semantic segmentation challenge for land cover description using multimodal remote sensing earth observation data, delving into real-world scenarios with a dataset comprising 70,000+ aerial imagery patches and 50,000 Sentinel-2 satellite acquisitions.
This library provides packages on DoubleML / Causal Machine Learning and Neural Networks in Python for Simulation and Case Studies.
Utilizing a multimodal architecture to predict the appropriate speaker turn in a dialogue.
A multi modal pipeline to generate three tones of reviews [harsh, constructive, kind] for a given artwork using fine-tuned Flan-T5 models.
This project use OCR for extracting prescription information and searching for related pharmacy
This is the official repository for Vista dataset - A Vietnamese multimodal dataset contains more than 700,000 samples of conversations and images
Omni-Modality Processing, Understanding, and Generation
Visual Instruction Tuning for Qwen2 Base Model
An intentionally simple Image to Food cross-modal search. Created by Prithiviraj Damodaran.
AMR extension for the spatial domain, with grounded frame of reference tracking
AnyModality is an open-source library to simplify MultiModal LLM inference and deployment.
Add a description, image, and links to the multimodal topic page so that developers can more easily learn about it.
To associate your repository with the multimodal topic, visit your repo's landing page and select "manage topics."