furkankupcu's Stars
rasbt/LLMs-from-scratch
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
karpathy/LLM101n
LLM101n: Let's build a Storyteller
KwaiVGI/LivePortrait
Bring portraits to life!
facebookresearch/segment-anything-2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
firmai/industry-machine-learning
A curated list of applied machine learning and data science notebooks and libraries across different industries (by @firmai)
ltdrdata/ComfyUI-Manager
ComfyUI-Manager is an extension designed to enhance the usability of ComfyUI. It offers management functions to install, remove, disable, and enable various custom nodes of ComfyUI. Furthermore, this extension provides a hub feature and convenience functions to access a wide range of information within ComfyUI.
AbdullahAlfaraj/Auto-Photoshop-StableDiffusion-Plugin
A user-friendly plug-in that makes it easy to generate stable diffusion images inside Photoshop using either Automatic or ComfyUI as a backend.
timothybrooks/instruct-pix2pix
jpmorganchase/python-training
Python training for business analysts and traders
roboflow/notebooks
Examples and tutorials on using SOTA computer vision models and techniques. Learn everything from old-school ResNet, through YOLO and object-detection transformers like DETR, to the latest models like Grounding DINO and SAM.
ZHO-ZHO-ZHO/ComfyUI-Workflows-ZHO
我的 ComfyUI 工作流合集 | My ComfyUI workflows collection
ifzhang/ByteTrack
[ECCV 2022] ByteTrack: Multi-Object Tracking by Associating Every Detection Box
cubiq/ComfyUI_IPAdapter_plus
Kosinkadink/ComfyUI-AnimateDiff-Evolved
Improved AnimateDiff for ComfyUI and Advanced Sampling Support
AcademySoftwareFoundation/openvdb
OpenVDB - Sparse volume data structure and tools
roboflow/sports
computer vision and sports
MrForExample/ComfyUI-3D-Pack
An extensive node suite that enables ComfyUI to process 3D inputs (Mesh & UV Texture, etc) using cutting edge algorithms (3DGS, NeRF, etc.)
FurkanGozukara/Stable-Diffusion
FLUX, Stable Diffusion, SDXL, SD3, LoRA, Fine Tuning, DreamBooth, Training, Automatic1111, Forge WebUI, SwarmUI, DeepFake, TTS, Animation, Text To Video, Tutorials, Guides, Lectures, Courses, ComfyUI, Google Colab, RunPod, Kaggle, NoteBooks, ControlNet, TTS, Voice Cloning, AI, AI News, ML, ML News, News, Tech, Tech News, Kohya, Midjourney, RunPod
isaac-sim/IsaacGymEnvs
Isaac Gym Reinforcement Learning Environments
fofr/cog-face-to-many
Turn any face into a video game character, pixel art, claymation, 3D or toy
WenjieDu/PyPOTS
A Python toolkit/library for reality-centric machine/deep learning and data mining on partially-observed time series, including SOTA neural network models for scientific analysis tasks of imputation/classification/clustering/forecasting/anomaly detection/cleaning on incomplete industrial (irregularly-sampled) multivariate TS with NaN missing values
kijai/ComfyUI-KJNodes
Various custom nodes for ComfyUI
AkshitIreddy/Interactive-LLM-Powered-NPCs
Interactive LLM Powered NPCs, is an open-source project that completely transforms your interaction with non-player characters (NPCs) in any game! 🎮🤖🚀
Unity-Technologies/ui-toolkit-manual-code-examples
Unity UI Toolkit documentation code examples
cloneofsimo/minRF
Minimal implementation of scalable rectified flow transformers, based on SD3's approach
git-disl/awesome-LLM-game-agent-papers
A Survey on Large Language Model-Based Game Agents
dimtoneff/ComfyUI-PixelArt-Detector
Generate, downscale, change palletes and restore pixel art images with SDXL.
facebookresearch/3D-Vision-and-Touch
When told to understand the shape of a new object, the most instinctual approach is to pick it up and inspect it with your hand and eyes in tandem. Here, touch provides high fidelity localized information while vision provides complementary global context. However, in 3D shape reconstruction, the complementary fusion of visual and haptic modalities remains largely unexplored. In this paper, we study this problem and present an effective chart-based approach to fusing vision and touch, which leverages advances in graph convolutional networks. To do so, we introduce a dataset of simulated touch and vision signals from the interaction between a robotic hand and a large array of 3D objects. Our results show that (1) leveraging both vision and touch signals consistently improves single-modality baselines, especially when the object is occluded by the hand touching it; (2) our approach outperforms alternative modality fusion methods and strongly benefits from the proposed chart-based structure; (3) reconstruction quality boosts with the number of grasps provided; and (4) the touch information not only enhances the reconstruction at the touch site but also extrapolates to its local neighborhood.
mikkel/ComfyUI-text-overlay
Overlay text on an image in ComfyUI with font/alignment/placement customization
VICO-UoE/OddOneOutAD