{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"LFToolbox-CLIM_VSENSE","owner":"V-Sense","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"MATLAB","color":"#e16737"},"pullRequestCount":1,"issueCount":1,"starsCount":7,"forksCount":2,"license":"BSD 2-Clause \"Simplified\" License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-22T17:47:37.853Z"}},{"type":"Public","name":"AutoMeshTracker","owner":"V-Sense","isFork":false,"description":"Offical repo for \"Moynihan, M., Ruano, S., Pagés, R. and Smolic, A., 2021. Autonomous Tracking For Volumetric Video Sequences\"","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":4,"starsCount":30,"forksCount":5,"license":"GNU General Public License v2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-05-06T15:20:48.290Z"}},{"type":"Public","name":"ACTION-Net","owner":"V-Sense","isFork":false,"description":"Official PyTorch implementation of ACTION-Net: Multipath Excitation for Action Recognition (CVPR'21)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":10,"starsCount":196,"forksCount":45,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-04-19T22:24:33.895Z"}},{"type":"Public","name":"Deep-Color-Mismatch-Correction","owner":"V-Sense","isFork":false,"description":"Deep learning-based solution for the correction of color mismatch in stereoscopic 3D images.","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-02-02T11:03:15.314Z"}},{"type":"Public","name":"AutoSkeleton","owner":"V-Sense","isFork":false,"description":"Trinity College Dublin developed auto-skeleton tools that can be used to quickly create skeletal structures from human character meshes in the Houdini 3D animation software.","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":66,"forksCount":6,"license":"GNU General Public License v3.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-12-14T00:35:31.266Z"}},{"type":"Public","name":"LFBM5D","owner":"V-Sense","isFork":false,"description":"C/C++ implementation of the LFBM5D filter for light field denoising and super-resolution. https://v-sense.scss.tcd.ie/?p=893","allTopics":["super-resolution","denoising","light-field"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":27,"forksCount":10,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-11-09T13:56:32.549Z"}},{"type":"Public","name":"DeepNormals","owner":"V-Sense","isFork":false,"description":"Code and Dataset from Deep Normal Estimation for Automatic Shading of Hand-Drawn Characters","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":6,"starsCount":180,"forksCount":19,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-09-21T12:37:28.202Z"}},{"type":"Public","name":"VI_VMAF_4_360","owner":"V-Sense","isFork":true,"description":"Helper scripts for the paper of Voronoi-based Objective Quality Metrics for Omnidirectional Video","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":4,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-06-09T23:34:22.380Z"}},{"type":"Public","name":"soft_segmentation","owner":"V-Sense","isFork":false,"description":"Our implementation of the paper 'Unmixing-Based Soft Color Segmentation for Image Manipulation'","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":15,"forksCount":7,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2020-02-09T12:28:32.758Z"}},{"type":"Public","name":"VR_user_behaviour","owner":"V-Sense","isFork":false,"description":"This is the GitHub page for the ACM TOMM 2020 paper: Do Users Behave Similarly in VR? Investigation of the Influence on the System Design","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2019-12-15T15:10:11.190Z"}},{"type":"Public","name":"Aesthetic-Image-Captioning-ICCVW-2019","owner":"V-Sense","isFork":false,"description":"Code and data for Aesthetic Image Captioning from Weakly-Labelled Photographs","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":31,"forksCount":7,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2019-10-24T13:22:01.668Z"}},{"type":"Public","name":"resources","owner":"V-Sense","isFork":true,"description":"A list of data sets and other resources for light fields for computer vision","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":77,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2019-10-21T09:36:58.936Z"}},{"type":"Public","name":"A-Geometry-Sensitive-Approach-for-Photographic-Style-Classification","owner":"V-Sense","isFork":false,"description":"IMVIP 2018","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":7,"forksCount":3,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2019-10-17T17:34:25.145Z"}},{"type":"Public","name":"ssar","owner":"V-Sense","isFork":false,"description":"Network for Ego Gesture Recogntion","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":5,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2019-09-18T16:18:35.111Z"}},{"type":"Public","name":"360SR","owner":"V-Sense","isFork":false,"description":"This repository contains python implementation of the paper: Super-resolution of Omnidirectional Images Using Adversarial Learning","allTopics":["vr","gan","super-resolution","360-video"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":20,"forksCount":9,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2019-09-13T13:33:20.035Z"}},{"type":"Public","name":"colornet-estimating-colorfulness","owner":"V-Sense","isFork":false,"description":"ColorNet: A learning-based colorfulness estimator for natural images","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":15,"forksCount":3,"license":"GNU General Public License v3.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2019-09-11T14:51:47.715Z"}},{"type":"Public","name":"360AudioVisual","owner":"V-Sense","isFork":false,"description":"This repository contains materials for the paper: Towards generating ambisonics using audio-visual cue for virtual reality","allTopics":["vr","dataset","ambisonics","360-video"],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":10,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2019-07-02T14:45:55.057Z"}},{"type":"Public","name":"CPM_PF","owner":"V-Sense","isFork":false,"description":"Coarse to fine Patch Match + Permeability Filter","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":6,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2019-06-19T10:57:02.107Z"}},{"type":"Public","name":"LFToolbox_Recolouring_HPR","owner":"V-Sense","isFork":false,"description":"Recolouring step of the Light Field toolbox + hot pixel removal","allTopics":[],"primaryLanguage":{"name":"MATLAB","color":"#e16737"},"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":3,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2018-11-13T14:01:12.811Z"}},{"type":"Public","name":"EgoCentricGestureNet","owner":"V-Sense","isFork":false,"description":"A repo with model and links to dataset from Egocentric Gesture Recognition for Head-Mounted AR devices (ISMAR 2018 Adjunct)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":7,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2018-09-20T09:59:17.466Z"}},{"type":"Public","name":"salnet360","owner":"V-Sense","isFork":false,"description":"Repository for implementation of SalNet360 in Caffe","allTopics":["caffe","deep-learning","saliency-map","360-photo","saliency"],"primaryLanguage":{"name":"MATLAB","color":"#e16737"},"pullRequestCount":0,"issueCount":2,"starsCount":18,"forksCount":3,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2018-07-05T18:14:01.271Z"}},{"type":"Public","name":"omniAttention","owner":"V-Sense","isFork":true,"description":"This is the GitHub page for the QoMEX 2018 paper: Visual Attention in Omnidirectional Video for Virtual Reality Applications","allTopics":["virtual-reality","attention","360-video","saliency-map","saliency","attention-model","saliency-detection"],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":6,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2018-05-24T16:14:50.771Z"}}],"repositoryCount":22,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}