-
ImVisionLabs Inc.
- Tokyo, Japan
-
13:32
(UTC +09:00) - www.linkedin.com/in/kenta-itakura-b88129202
- @imvisionlabs
- https://imvisionlabs.com/
Block or Report
Block or report KentaItakura
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseStars
Language
Sort by: Recently starred
3D Rendering Toolbox: Color images and depth maps from 3D meshes.
LLaVA-JP is a Japanese VLM trained by LLaVA method
A full Python Implementation of the ROUGE Metric (not a wrapper)
ROS/ROS2 enabled Sensor models (Assets) on Unity
[NeurIPS2023] Code release for "Hierarchical Open-vocabulary Universal Image Segmentation"
This is the python interface for adapted COLMAP; you can run both SfM and MVS.
Connect MATLAB to LLM APIs, including OpenAI® Chat Completions, Azure® OpenAI Services, and Ollama™
[CVPR 2024] Spacetime Gaussian Feature Splatting for Real-Time Dynamic View Synthesis
This is a PyTorch implementation of the ECCV2020 paper "DeepSFM: Structure From Motion Via Deep Bundle Adjustment".
livox_camera_lidar_calibrationの日本語ドキュメント
Segment Anything in 3D with NeRFs (NeurIPS 2023)
[CVPR 2024] 4D Gaussian Splatting for Real-Time Dynamic Scene Rendering
High-Resolution Multispectral Dataset for Semantic Segmentation
Project Page for "LISA: Reasoning Segmentation via Large Language Model"
Python package for segmenting LiDAR data using Segment-Anything Model (SAM) from Meta AI.
✨✨Latest Advances on Multimodal Large Language Models
Official implementation of "Neuralangelo: High-Fidelity Neural Surface Reconstruction" (CVPR 2023)
[ICCV'21] UNISURF: Unifying Neural Implicit Surfaces and Radiance Fields for Multi-View Reconstruction
mPLUG-Owl & mPLUG-Owl2: Modularized Multimodal Large Language Model
OpenFace – a state-of-the art tool intended for facial landmark detection, head pose estimation, facial action unit recognition, and eye-gaze estimation.
Code for 3D-LLM: Injecting the 3D World into Large Language Models