Skip to content
forked from langgenius/dify

Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.

License

Notifications You must be signed in to change notification settings

frost-rolf/dify

Repository files navigation

English | 简体中文 | 日本語 | Español | Klingon | Français

Static Badge chat on Discord follow on Twitter Docker Pulls

📌 Check out Dify Premium on AWS and deploy it to your own AWS VPC with one-click.

Dify is an open-source LLM app development platform. Dify's intuitive interface combines a RAG pipeline, AI workflow orchestration, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.

workflow_intro.mp4

Using our Cloud Services

You can try out Dify.AI Cloud now. It provides all the capabilities of the self-deployed version, and includes 200 free GPT-4 calls.

Looking to purchase via AWS?

Check out Dify Premium on AWS and deploy it to your own AWS VPC with one-click.

Features

1. Workflow: Create and test complex AI workflows on a visual canvas, with pre-built nodes taking advantage of the power of all the following features and beyond.

2. Support: Seamless integration with hundreds of proprietary / open-source LLMs and dozens of inference providers, including GPT, Mistral, Llama2, and OpenAI API-compatible models. A full list of supported model providers is kept here.

3. Prompt IDE: Visual orchestration of applications and services based on any LLMs. Easily share with your team.

4. RAG Engine: Includes various RAG capabilities based on full-text indexing or vector database embeddings, allowing direct upload of PDFs, TXTs, and other text formats.

5. AI Agent: Based on Function Calling and ReAct, the Agent inference framework allows users to customize tools, what you see is what you get. Dify provides more than a dozen built-in tools for AI agents, such as Google Search, DELL·E, Stable Diffusion, WolframAlpha, etc.

6. LLMOps: Monitor and analyze application logs and performance, continuously improving Prompts, datasets, or models based on production data.

Dify vs. LangChain vs. Assistants API

Feature Dify.AI Assistants API LangChain
Programming Approach API-oriented API-oriented Python Code-oriented
Ecosystem Strategy Open Source Close Source Open Source
RAG Engine Supported Supported Not Supported
Prompt IDE Included Included None
Supported LLMs Rich Variety OpenAI-only Rich Variety
Local Deployment Supported Not Supported Not Applicable

Before You Start

Star us on GitHub, and be instantly notified for new releases! star-us

Install the Community Edition

System Requirements

Before installing Dify, make sure your machine meets the following minimum system requirements:

  • CPU >= 2 Core
  • RAM >= 4GB

Quick Start

The easiest way to start the Dify server is to run our docker-compose.yml file. Before running the installation command, make sure that Docker and Docker Compose are installed on your machine:

cd docker
docker compose up -d

After running, you can access the Dify dashboard in your browser at http://localhost/install and start the initialization installation process.

Deploy with Helm Chart

Helm Chart version, which allows Dify to be deployed on Kubernetes.

Configuration

If you need to customize the configuration, please refer to the comments in our docker-compose.yml file and manually set the environment configuration. After making the changes, please run docker-compose up -d again. You can see the full list of environment variables in our docs.

Star History

Star History Chart

Contributing

For those who'd like to contribute code, see our Contribution Guide.

At the same time, please consider supporting Dify by sharing it on social media and at events and conferences.

Projects made by community

Contributors

Translations

We are looking for contributors to help with translating Dify to languages other than Mandarin or English. If you are interested in helping, please see the i18n README for more information, and leave us a comment in the global-users channel of our Discord Community Server.

Community & Support

  • Github Discussion. Best for: sharing feedback and checking out our feature roadmap.
  • GitHub Issues. Best for: bugs you encounter using Dify.AI, and feature proposals. See our Contribution Guide.
  • Email Support. Best for: questions you have about using Dify.AI.
  • Discord. Best for: sharing your applications and hanging out with the community.
  • Twitter. Best for: sharing your applications and hanging out with the community.
  • Business Contact. Best for: business inquiries of licensing Dify.AI for commercial use.

Direct Meetings

Help us make Dify better. Reach out directly to us.

Point of Contact Purpose
Git-Hub-README-Button-3x Product design feedback, user experience discussions, feature planning and roadmaps.
Git-Hub-README-Button-2x Technical support, issues, or feature requests

Security Disclosure

To protect your privacy, please avoid posting security issues on GitHub. Instead, send your questions to [email protected] and we will provide you with a more detailed answer.

License

This repository is available under the Dify Open Source License, which is essentially Apache 2.0 with a few additional restrictions.

About

Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • TypeScript 49.6%
  • Python 45.5%
  • MDX 2.8%
  • CSS 1.3%
  • JavaScript 0.4%
  • SCSS 0.2%
  • Other 0.2%