Stars
DeepSeek LLM: Let there be answers
这是一个用于显示当前网速、CPU及内存利用率的桌面悬浮窗软件,并支持任务栏显示,支持更换皮肤。
PyTorch implementation of DQN, AC, ACER, A2C, A3C, PG, DDPG, TRPO, PPO, SAC, TD3 and ....
PyTorch implementations of deep reinforcement learning algorithms and environments
A toolkit for developing and comparing reinforcement learning algorithms.
Implementation of Reinforcement Learning Algorithms. Python, OpenAI Gym, Tensorflow. Exercises and Solutions to accompany Sutton's Book and David Silver's course.
Simple Reinforcement learning tutorials, 莫烦Python 中文AI教学
Global search algorithms for finding optimal tensor network contraction sequences.
This is the homepage of a new book entitled "Mathematical Foundations of Reinforcement Learning."
🚀 「大模型」3小时从0训练27M参数的视觉多模态VLM!🌏 Train a 27M-parameter VLM from scratch in just 3 hours!
Transformer Explained Visually: Learn How LLM Transformer Models Work with Interactive Visualization
High-Resolution Image Synthesis with Latent Diffusion Models
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use th…
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
An Autonomous LLM Agent for Complex Task Solving
Create Customized Software using Natural Language Idea (through LLM-powered Multi-Agent Collaboration)
A programming framework for agentic AI 🤖 PyPi: autogen-agentchat Discord: https://aka.ms/autogen-discord Office Hour: https://aka.ms/autogen-officehour
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
🚀🚀 「大模型」2小时完全从0训练26M的小参数GPT!🌏 Train a 26M-parameter GPT from scratch in just 2h!
Google AI 2018 BERT pytorch implementation
Easy and Efficient Finetuning LLMs. (Supported LLama, LLama2, LLama3, Qwen, Baichuan, GLM , Falcon) 大模型高效量化训练+部署.
Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
An Open-source Neural Hierarchical Multi-label Text Classification Toolkit