Highlights
- Pro
Stars
When it comes to optimizers, it's always better to be safe than sorry
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga…
A PyTorch library for implementing flow matching algorithms, featuring continuous and discrete flow matching implementations. It includes practical examples for both text and image modalities.
ReFT: Representation Finetuning for Language Models
MTEB: Massive Text Embedding Benchmark
You See it, You Got it: Learning 3D Creation on Pose-Free Videos at Scale
WhisperX: Automatic Speech Recognition with Word-level Timestamps (& Diarization)
DINO-X: The World's Top-Performing Vision Model for Open-World Object Detection and Understanding
HunyuanVideo: A Systematic Framework For Large Video Generation Model
Instruct-tune LLaMA on consumer hardware
Code for 'LLM2Vec: Large Language Models Are Secretly Powerful Text Encoders'
[EMNLP 2021] SimCSE: Simple Contrastive Learning of Sentence Embeddings https://arxiv.org/abs/2104.08821
This repo contains the code and data for "VLM2Vec: Training Vision-Language Models for Massive Multimodal Embedding Tasks"
This repository provides the code and model checkpoints for AIMv1 and AIMv2 research projects.
✨✨Latest Advances on Multimodal Large Language Models
LLM2CLIP makes SOTA pretrained CLIP model more SOTA ever.
Efficient Multimodal Large Language Models: A Survey
A trainable PyTorch reproduction of AlphaFold 3.
Official Repo for Paper "OmniEdit: Building Image Editing Generalist Models Through Specialist Supervision"
Finetune Llama 3.3, Mistral, Phi, Qwen 2.5 & Gemma LLMs 2-5x faster with 80% less memory
各种安全相关思维导图整理收集。渗透步骤,web安全,CTF,业务安全,人工智能,区块链安全,数据安全,安全开发,无线安全,社会工程学,二进制安全,移动安全,红蓝对抗,运维安全,风控安全,linux安全
为GPT/GLM等LLM大语言模型提供实用化交互接口,特别优化论文阅读/润色/写作体验,模块化设计,支持自定义快捷按钮&函数插件,支持Python和C++等项目剖析&自译解功能,PDF/LaTex论文翻译&总结功能,支持并行问询多种LLM模型,支持chatglm3等本地模型。接入通义千问, deepseekcoder, 讯飞星火, 文心一言, llama2, rwkv, claude2, m…
Mind the Gap: Understanding the Modality Gap in Multi-modal Contrastive Representation Learning
Drawing Bayesian networks, graphical models, tensors, technical frameworks, and illustrations in LaTeX.
MLCD & UNICOM : Large-Scale Visual Representation Model
Efficient Triton Kernels for LLM Training