Compare commits

..

879 Commits
0.4.0 ... 0.7.0

Author SHA1 Message Date
Alex
2be523cf77 Fix handling of embeddings_key in api_search() function 2024-01-30 17:22:33 +00:00
Alex
c01e334487 Merge pull request #843 from larinam/fix_application
Fix application + script requirements.txt
2024-01-29 20:59:36 +00:00
Alex
a2418d1373 Add sentence-transformers library to requirements.txt and comment out model_name in base.py 2024-01-29 20:51:28 +00:00
Alex
a697248b26 Merge pull request #841 from ManishMadan2882/main
Message bubble responsiveness
2024-01-29 13:55:29 +00:00
ManishMadan2882
6058939c00 change size in copy, like , dislike icons 2024-01-29 19:10:03 +05:30
Anton Larin
318de530e3 fix openapi-parser requirement 2024-01-27 16:52:33 +01:00
Anton Larin
9e04b7796a application folder related changes:
* optimize content of requirements.txt
* upgrade libs
* fix imports
2024-01-27 16:25:19 +01:00
Anton Larin
e8099c4db5 script folder related changes:
* optmize content of requirements.txt
* upgrade libs
* fix imports
2024-01-27 14:58:08 +01:00
Alex
bf808811cc Update README.md 2024-01-26 12:21:09 +00:00
ManishMadan2882
f0293de1b9 ui adjustments 2024-01-26 03:06:15 +05:30
ManishMadan2882
810dcb90ce refactored the divs, prevent overlap 2024-01-26 02:47:51 +05:30
ManishMadan2882
a2f2b8fabc make responsive msg bubble 2024-01-26 02:33:50 +05:30
Alex
cbc5f47786 Merge pull request #837 from ManishMadan2882/main
Adding Dark Mode
2024-01-23 14:59:22 +00:00
ManishMadan2882
3e3886ced7 slight UI changes 2024-01-23 19:38:22 +05:30
ManishMadan2882
9ce39fd2ba made borders in settings a bit darker 2024-01-22 18:22:46 +05:30
ManishMadan2882
5b08cdedf0 revert changes in docker yaml 2024-01-22 16:22:19 +05:30
ManishMadan2882
67e4d40c49 added dark mode, About page and bubble icons 2024-01-22 16:11:26 +05:30
ManishMadan2882
537a733157 add dark mode - conversation, bubble, UI fixes 2024-01-22 02:56:07 +05:30
ManishMadan2882
5136e7726d added dark mode, Hero component 2024-01-21 17:18:23 +05:30
Alex
6e236ba74d Merge pull request #827 from Juneezee/vite-5
Upgrade to Vite 5
2024-01-19 14:01:40 +00:00
Eng Zer Jun
374b665089 Upgrade to Vite 5
This commit upgrades vite to the latest version 5, and also updates the
vite plugins to the latest version.

Signed-off-by: Eng Zer Jun <engzerjun@gmail.com>
2024-01-19 21:34:28 +08:00
ManishMadan2882
ffecc9a0c7 add dark mode, in Settings 2024-01-19 19:01:13 +05:30
ManishMadan2882
0b997418d3 add dark - sidebar 2024-01-19 01:47:23 +05:30
ManishMadan2882
eaad8a4cf5 initialising dark mode 2024-01-18 02:39:40 +05:30
Alex
396b4595f4 Merge pull request #832 from ArnabBCA/main
Fixed Empty Document Name Upload
2024-01-16 17:07:02 +00:00
Arnab Ghosh
0752aae9ef Fixed Empty Document Name Upload 2024-01-16 15:35:48 +05:30
Alex
ad2221a677 Merge pull request #830 from arc53/feature/search-endpoint
Search docs not inside the /stream in the stream request
2024-01-15 16:55:53 +00:00
Alex
1713d693b1 Merge pull request #831 from ManishMadan2882/feature/search-endpoint
integrate /api/search endpoint, get sources post stream
2024-01-15 16:34:48 +00:00
ManishMadan2882
f4f056449f integrate /api/search endpoint, get sources post stream 2024-01-15 20:23:18 +05:30
Alex
6a70e3e45b Commented out unused code in api_search function 2024-01-12 14:39:17 +00:00
Alex
a04cdee33f Refactor source log generation in complete_stream function 2024-01-12 14:38:15 +00:00
Alex
157769eeb4 Add API endpoint for searching documents 2024-01-12 14:35:23 +00:00
Alex
667b66b926 Merge pull request #825 from arc53/feat/mongodb
Public LLM
2024-01-09 14:31:40 +00:00
Alex
c0f7b344d9 Update environment variables and installation instructions 2024-01-09 12:35:18 +00:00
Alex
060c59e97d Update mpnet-base-v2.zip download URL 2024-01-09 11:41:25 +00:00
Alex
b3461b7134 Add MPNet model and update vector store for Hugging Face embeddings 2024-01-09 11:39:32 +00:00
Pavel
001c450abb choice text 2024-01-09 13:05:16 +03:00
Pavel
ceaa5763d4 choice fix 2024-01-09 12:57:19 +03:00
Alex
b45fd58944 Update EMBEDDINGS_NAME in settings.py and test_vector_store.py 2024-01-09 00:34:04 +00:00
Alex
b3149def82 Update EMBEDDINGS_NAME in settings.py 2024-01-09 00:29:02 +00:00
Alex
378d498402 Remove unused imports in docsgpt_provider.py 2024-01-09 00:19:49 +00:00
Alex
98f52b32a3 Update README and Quickstart guide 2024-01-09 00:18:04 +00:00
Alex
0ab32a6f84 Update setup.sh script with new options for language model usage 2024-01-09 00:07:37 +00:00
Alex
71cc22325d Add application files and update setup script 2024-01-09 00:05:44 +00:00
Alex
e1b2991aa6 Update LLM_NAME and EMBEDDINGS_NAME 2024-01-09 00:01:31 +00:00
Alex
033bcf80d0 docsgpt llm provider 2024-01-08 23:35:37 +00:00
Alex
103118d558 Merge pull request #823 from ManishMadan2882/main
fix distortion on different browsers
2024-01-08 11:30:30 +00:00
ManishMadan2882
f91b5fa004 fix distortion on different browsers 2024-01-08 16:03:41 +05:30
Alex
7179bf7b67 Merge pull request #822 from arc53/feat/mongodb
Mongodb integration as vectorstore
2024-01-06 18:30:25 +00:00
Alex
a3e6239e6e fix: remove import 2024-01-06 18:23:20 +00:00
Alex
1fa12e56c6 Remove unused test cases in test_openai.py 2024-01-06 18:04:50 +00:00
Alex
4ff834de76 Refactor MongoDBVectorStore and add delete_index method 2024-01-06 17:59:01 +00:00
Alex
293b7b09a9 init tests 2024-01-05 17:16:16 +00:00
Alex
d5945f9ee7 Update README.md 2024-01-05 13:58:22 +00:00
Alex
d1f5a6fc31 Merge pull request #816 from ManishMadan2882/main
adding responsive sidebar
2024-01-04 20:51:30 +00:00
ManishMadan2882
e7b9f5e4c3 adding responsive sidebar 2024-01-05 01:50:52 +05:30
Alex
7870749077 fix openai 2024-01-03 12:09:05 +00:00
Alex
c5352f443a Merge pull request #813 from CBID2/making-alt-text-less-redundant
fix: Making alt text less redundant
2024-01-03 10:52:08 +00:00
Christine
fd8b7aa0f2 fix: change alt text for setting 2024-01-03 04:17:53 +00:00
Christine
458ea266ec fix: change name to alt text for Discord and GitHub 2024-01-03 04:14:06 +00:00
Alex
9748eaba25 Merge pull request #811 from Rutam21/patch-2
Added new Deployment Guide for Kamatera Performance Cloud
2023-12-31 15:37:32 +00:00
Alex
887a3740b2 Update holopin.yml 2023-12-31 15:34:55 +00:00
Rutam Prita Mishra
2e7cfe9cd7 Added new Deployment Guide
This PR adds a new deployment guide for Kamatera Performance Cloud.
2023-12-26 16:28:57 +05:30
Alex
6dbe156a02 Update README.md 2023-12-25 11:52:37 +00:00
Alex
2a9ef6d48e Merge pull request #792 from arc53/dependabot/npm_and_yarn/frontend/vite-4.5.1
Bump vite from 4.5.0 to 4.5.1 in /frontend
2023-12-22 15:51:11 +00:00
Alex
6717ddbd0b Merge pull request #804 from arc53/dependabot/npm_and_yarn/docs/next-13.5.1
Bump next from 13.4.19 to 13.5.1 in /docs
2023-12-22 15:50:49 +00:00
Alex
47c1aab064 Merge pull request #793 from arc53/dependabot/npm_and_yarn/extensions/react-widget/vite-4.4.12
Bump vite from 4.4.9 to 4.4.12 in /extensions/react-widget
2023-12-22 15:50:08 +00:00
Alex
eda41658b9 Merge pull request #806 from arc53/cve/py-removal
fix: Remove py==1.11.0 from requirements.txt
2023-12-22 15:47:45 +00:00
Alex
7f79363944 fix: Remove py==1.11.0 from requirements.txt 2023-12-22 15:44:39 +00:00
Alex
25967f2a09 Merge pull request #805 from arc53/fix/scripts-vulnerabilities
fix: vulns
2023-12-22 15:35:37 +00:00
Alex
4d3963ad67 fix: vulns 2023-12-22 15:27:23 +00:00
dependabot[bot]
f78c5257dc Bump next from 13.4.19 to 13.5.1 in /docs
Bumps [next](https://github.com/vercel/next.js) from 13.4.19 to 13.5.1.
- [Release notes](https://github.com/vercel/next.js/releases)
- [Changelog](https://github.com/vercel/next.js/blob/canary/release.js)
- [Commits](https://github.com/vercel/next.js/compare/v13.4.19...v13.5.1)

---
updated-dependencies:
- dependency-name: next
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-12-22 14:14:27 +00:00
Alex
ccc6234ac8 Merge pull request #803 from arc53/fix/dependency-upgrades
fix: cve upgrades
2023-12-22 14:13:41 +00:00
Alex
c81b0200eb fix: Update pydantic_settings version to 2.1.0 2023-12-22 14:08:03 +00:00
Alex
f039d37c8a fix: pydantic 2023-12-22 14:03:43 +00:00
Alex
237975bfef fix: cve upgrades 2023-12-22 13:25:57 +00:00
Alex
015bc7c8c3 hotfix source doc data 2023-12-22 12:10:35 +00:00
Alex
3da2a00ee9 Merge pull request #801 from Victorivus/bug/fix-#800-mistral_models
Update requirements.txt HF Transformers
2023-12-14 15:29:28 +00:00
Victorivus
16eca5bebf Update requirements.txt HF Transformers
Fix 'mistral' models missing
2023-12-12 15:16:31 +01:00
Alex
a4483cf255 Revert "Merge pull request #797 from arc53/fix_boxes_hero"
This reverts commit 0bf020a1b4, reversing
changes made to a62566e8fb.
2023-12-11 12:18:47 +00:00
Alex
0bf020a1b4 Merge pull request #797 from arc53/fix_boxes_hero
Fix_hero
2023-12-10 19:17:46 +00:00
Pavel
d43927a167 Fix_hero
Fix of hero section for chrome and firefox. Fix of conversations-container pushing containers below.
2023-12-10 16:28:20 +03:00
Alex
a62566e8fb Merge pull request #795 from HeetVekariya/fix/UI
fix: API docs text overflow
2023-12-08 17:08:43 +00:00
HeetVekariya
23a1730106 fix: API docs text overflow 2023-12-08 20:36:12 +05:30
dependabot[bot]
f8ac5e0af3 Bump vite from 4.4.9 to 4.4.12 in /extensions/react-widget
Bumps [vite](https://github.com/vitejs/vite/tree/HEAD/packages/vite) from 4.4.9 to 4.4.12.
- [Release notes](https://github.com/vitejs/vite/releases)
- [Changelog](https://github.com/vitejs/vite/blob/v4.4.12/packages/vite/CHANGELOG.md)
- [Commits](https://github.com/vitejs/vite/commits/v4.4.12/packages/vite)

---
updated-dependencies:
- dependency-name: vite
  dependency-type: direct:development
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-12-06 01:52:56 +00:00
dependabot[bot]
eb48a153d9 Bump vite from 4.5.0 to 4.5.1 in /frontend
Bumps [vite](https://github.com/vitejs/vite/tree/HEAD/packages/vite) from 4.5.0 to 4.5.1.
- [Release notes](https://github.com/vitejs/vite/releases)
- [Changelog](https://github.com/vitejs/vite/blob/v4.5.1/packages/vite/CHANGELOG.md)
- [Commits](https://github.com/vitejs/vite/commits/v4.5.1/packages/vite)

---
updated-dependencies:
- dependency-name: vite
  dependency-type: direct:development
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-12-05 23:44:40 +00:00
Alex
1a78a6f786 Merge pull request #791 from arc53/bug/sticky-prompts
Add getLocalPrompt and setLocalPrompt functions to preferenceApi.ts
2023-12-04 11:27:18 +02:00
Alex
f8f60c62fe Add getLocalPrompt and setLocalPrompt functions to preferenceApi.ts 2023-12-04 11:23:51 +02:00
Alex
453e507b89 Update routes.py 2023-11-24 15:17:48 +00:00
Alex
022c32f9d5 fix chats display 2023-11-24 12:22:34 +00:00
Alex
af1a0c3520 Update routes.py 2023-11-23 11:31:07 +00:00
Alex
0a6d9dfcf4 Merge pull request #782 from arc53/feature/add-prompts
Feature/add prompts
2023-11-23 00:13:02 +00:00
Alex
5bdedacab1 fix xss 2023-11-23 00:09:17 +00:00
Alex
d7a1be2f3c fix bug 2023-11-23 00:00:08 +00:00
Alex
d6dcbb63d4 fix ruff 2023-11-22 23:57:47 +00:00
Alex
b2770f67a1 custom prompts 2023-11-22 23:55:41 +00:00
Alex
aa2691b153 Merge pull request #513 from akshay11298/mock-backend-server
Mock backend server
2023-11-22 12:27:10 +00:00
Alex
e9a9cbbd07 feedback local 2023-11-22 12:16:37 +00:00
Alex
17e2222802 Merge pull request #777 from arc53/dependabot/pip/scripts/aiohttp-3.8.6
Bump aiohttp from 3.8.5 to 3.8.6 in /scripts
2023-11-22 09:30:43 +00:00
Alex
58b2970b19 Merge pull request #778 from arc53/dependabot/pip/application/aiohttp-3.8.6
Bump aiohttp from 3.8.5 to 3.8.6 in /application
2023-11-22 09:28:20 +00:00
Alex
fd69961185 Merge pull request #780 from guspan-tanadi/docsAPIdeleteold
docs: refer delete_old sample docs_check comments API-docs
2023-11-22 00:19:41 +00:00
Guspan Tanadi
e5cd813958 docs: delete_old sample docs_check comments API-docs 2023-11-21 22:15:27 +07:00
Alex
5b12423d98 setup-fix2 2023-11-21 10:16:54 +00:00
Alex
4141f633a3 Setup process 2023-11-21 10:16:10 +00:00
Alex
67854b3ebd Reqs2 2023-11-17 16:18:28 +00:00
Alex
0c21dbc7c8 reqs cmd 2023-11-17 16:18:22 +00:00
Alex
5925aa50d8 Merge pull request #779 from arc53/bug/fix-patj
fix path bug on default
2023-11-17 15:38:12 +00:00
Alex
852b016111 fix path bug on default 2023-11-17 15:33:51 +00:00
Alex
ba77a67ba7 fix path 2023-11-17 15:31:53 +00:00
Alex
c14a9a55d7 Merge pull request #775 from guspan-tanadi/markdownhighlightNote
style(README): markdown highlight Note section
2023-11-15 11:32:24 +00:00
dependabot[bot]
5203db6c9c Bump aiohttp from 3.8.5 to 3.8.6 in /application
Bumps [aiohttp](https://github.com/aio-libs/aiohttp) from 3.8.5 to 3.8.6.
- [Release notes](https://github.com/aio-libs/aiohttp/releases)
- [Changelog](https://github.com/aio-libs/aiohttp/blob/master/CHANGES.rst)
- [Commits](https://github.com/aio-libs/aiohttp/compare/v3.8.5...v3.8.6)

---
updated-dependencies:
- dependency-name: aiohttp
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-11-14 23:45:14 +00:00
dependabot[bot]
30eb8dda1d Bump aiohttp from 3.8.5 to 3.8.6 in /scripts
Bumps [aiohttp](https://github.com/aio-libs/aiohttp) from 3.8.5 to 3.8.6.
- [Release notes](https://github.com/aio-libs/aiohttp/releases)
- [Changelog](https://github.com/aio-libs/aiohttp/blob/master/CHANGES.rst)
- [Commits](https://github.com/aio-libs/aiohttp/compare/v3.8.5...v3.8.6)

---
updated-dependencies:
- dependency-name: aiohttp
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-11-14 23:42:27 +00:00
Guspan Tanadi
69d40b5fe8 Merge branch 'arc53:main' into markdownhighlightNote 2023-11-15 06:35:04 +07:00
Alex
706e87659e Merge pull request #776 from arc53/feature/settings-api
Feature/settings api
2023-11-14 01:21:37 +00:00
Alex
5c785e49af prompts and docs 2023-11-14 01:16:42 +00:00
Alex
0974085c6f prompts 2023-11-14 01:16:06 +00:00
Guspan Tanadi
e67ced8848 new line Important section Quickstart 2023-11-13 20:02:07 +07:00
Guspan Tanadi
c2dea6b881 bold colon Important section Quickstart 2023-11-13 19:58:35 +07:00
Guspan Tanadi
ee62b2cf31 lowercase Important section Quickstart 2023-11-13 19:56:23 +07:00
Guspan Tanadi
252e06bee6 revert markdown highlight Note Important section Quickstart 2023-11-13 19:54:12 +07:00
Guspan Tanadi
1f0ce88e08 style: highlight markdown Note Important section Quickstart 2023-11-13 19:33:01 +07:00
Guspan Tanadi
7e8fb388a3 style(README): highlight markdown available Note section 2023-11-13 19:17:13 +07:00
Alex
a3de360878 fix sidebar 2023-11-12 18:31:29 +00:00
Alex
6ee556e386 Delete HACKTOBERFEST.md 2023-11-12 18:26:14 +00:00
Alex
623ed89100 Update How-to-use-different-LLM.md 2023-11-08 14:34:33 +00:00
Alex
8c114cae95 Update How-to-use-different-LLM.md 2023-11-08 14:33:11 +00:00
Alex
93dd58ec59 Update How-to-use-different-LLM.md 2023-11-08 14:32:18 +00:00
Alex
f0bc93ad8e Update How-to-use-different-LLM.md 2023-11-08 14:30:53 +00:00
Alex
27e8aad479 Update HACKTOBERFEST.md 2023-11-08 11:41:02 +00:00
Alex
6298578db9 Merge pull request #764 from Rutam21/patch-3
Add DigitalOcean Droplet Hosting solution to Hosting-the-app.md
2023-11-07 20:29:51 +00:00
Alex
f079e5dadd Merge pull request #763 from Rutam21/patch-2
Corrected Typo on _meta.json
2023-11-07 20:29:29 +00:00
Alex
e6fdead89f Merge pull request #762 from SamsShow/code/setting-ui
[Code]:Setting Section UI
2023-11-07 18:11:44 +00:00
Alex
cfa6e3982c Update HACKTOBERFEST.md 2023-11-07 16:38:00 +00:00
Rutam Prita Mishra
e4bc4d9071 Add DigitalOcean Droplet Hosting solution to Hosting-the-app.md
This PR adds a new deployment solution for DocsGPT.
2023-11-06 19:29:51 +05:30
Rutam Prita Mishra
d372e10f1a Corrected Typo on _meta.json
This PR changes `Rainway` to the correct spelling `Railway`.
2023-11-06 19:22:58 +05:30
Saksham Tyagi
d1c93754db Merge branch 'arc53:main' into code/setting-ui 2023-11-04 02:38:09 +05:30
Saksham Tyagi
3d54a1abf3 widget 2023-11-04 02:33:56 +05:30
Saksham Tyagi
06eef5779d Documents 2023-11-03 22:52:16 +05:30
Alex
b4d78376fb Merge pull request #759 from krabbi/add_table_render
Render tables in conversation
2023-11-03 13:12:46 +00:00
Saksham Tyagi
87a59a6de3 Prompt 2023-11-03 13:06:18 +05:30
Alexander Deshkevich
1e7741e341 Render tables in conversation 2023-11-02 18:51:15 -03:00
Alex
ae5e484506 fix docx file 2023-11-02 18:25:43 +00:00
Alex
c9dd219565 Update CONTRIBUTING.md 2023-11-02 11:11:58 +00:00
Alex
55eb662dc9 Merge pull request #758 from arc53/bug/UI-fixes-and-settings-prep
Bug/UI fixes and settings prep
2023-11-01 22:10:04 +00:00
Alex
2d202088c7 syntax 2023-11-01 21:48:52 +00:00
Alex
9f7c9180d9 prep settings slightly 2023-11-01 21:45:40 +00:00
Alex
fdc5e0a92d Merge pull request #756 from ArpitPandey29/main
docs: fix grammar issues
2023-11-01 12:10:54 +00:00
arpitpandey0209
7f6fef1373 Revert txt files 2023-11-01 10:11:40 +05:30
ArpitPandey29
e4973f572f Update combine_prompt_hist.txt 2023-11-01 10:08:36 +05:30
ArpitPandey29
eb768f2076 Update combine_prompt.txt 2023-11-01 10:07:25 +05:30
ArpitPandey29
df9723a011 Update combine_prompt.txt 2023-11-01 10:06:54 +05:30
ArpitPandey29
51ad3fdb0b Update combine_prompt.txt 2023-11-01 10:04:53 +05:30
ArpitPandey29
8e553e7a93 Update combine_prompt.txt 2023-11-01 10:04:30 +05:30
ArpitPandey29
4c70e92293 Update chat_reduce_prompt.txt 2023-11-01 10:03:50 +05:30
ArpitPandey29
3e983b121e Update chat_combine_prompt.txt 2023-11-01 10:02:03 +05:30
ArpitPandey29
9f3c962ea4 Update chat_combine_prompt.txt 2023-11-01 10:01:43 +05:30
ArpitPandey29
2a1a3fb1b5 Update chat_combine_prompt.txt 2023-11-01 09:59:46 +05:30
ArpitPandey29
bb28cc5c65 Update README.md 2023-11-01 09:58:13 +05:30
ArpitPandey29
23b6a38e18 Update README.md 2023-11-01 09:57:31 +05:30
arpitpandey0209
715cd9daf5 Revert some files 2023-11-01 09:51:09 +05:30
ArpitPandey29
cbfdaec394 Merge branch 'arc53:main' into main 2023-11-01 09:45:39 +05:30
Alex
bb527ac981 Update index.mdx 2023-10-31 23:22:23 +00:00
Alex
961c26894d Update README.md 2023-10-31 23:21:46 +00:00
Alex
693bdebb30 Merge pull request #752 from varundhand/feature/mobile-resp
Feat: enhance mobile responsiveness of Hero Page final
2023-10-31 23:13:04 +00:00
Varun Dhand
353e24f1c5 fix: container layout for firefox 2023-11-01 03:56:48 +05:30
Alex
59d1773057 Merge pull request #757 from gfggithubleet/patch-1
Update README.md
2023-10-31 21:57:50 +00:00
Alex
93a1368b60 Merge pull request #754 from IamSenthilKumar/fix-docs
Fix Guide: How to use other LLMS
2023-10-31 21:48:14 +00:00
gfggithubleet
3bc0fe5a70 Update README.md
TYPO FIXED
2023-11-01 01:09:38 +05:30
arpitpandey0209
973c11a048 docs: fix grammar issues 2023-11-01 01:04:40 +05:30
Alex
5094386516 hotfix-1 2023-10-31 18:21:44 +00:00
Alex
64477c6573 Merge pull request #691 from SoumyadiptoPal/settings
Feature: Add "Settings" Button to sidebar
2023-10-31 18:20:58 +00:00
Alex
f052c707e7 Merge branch 'main' into settings 2023-10-31 18:18:31 +00:00
Senthil Kumar N
de0e1d3e10 Fix Guide to use other LLMS 2023-10-31 23:41:27 +05:30
Alex
e273da1b5b Merge pull request #753 from THEGAMECHANGER416/patch-1
Improved Documentation: Added comments in application/worker.py
2023-10-31 17:56:09 +00:00
Alex
761f6963ab Merge pull request #738 from 0xrahul6/docs/enhance
Enhanced Guides/Customization.md
2023-10-31 17:25:37 +00:00
Alex
e5aff1316a Merge pull request #739 from 0xrahul6/docs/trains
Enhancement: Updated Train other docs
2023-10-31 17:24:47 +00:00
Alex
8ee0fbe6a3 Update README.md 2023-10-31 17:23:01 +00:00
Arnav Kohli
4c6b8b4173 Update worker.py
Added comments in difficult to understand areas
2023-10-31 20:00:07 +05:30
Varun Dhand
6940a75591 erge branch 'feature/mobile-resp' of https://github.com/varundhand/DocsGPT into feature/mobile-resp 2023-10-31 19:33:03 +05:30
Varun Dhand
7ba939b008 feat: mobile responsive hero page 2023-10-31 19:27:45 +05:30
0xrahul6
6918a36bee Update Customising-prompts.md 2023-10-31 19:14:14 +05:30
0xrahul6
ba132fc411 Update Customising-prompts.md 2023-10-31 19:12:38 +05:30
Alex
b4a940a8d6 Merge pull request #750 from guspan-tanadi/settingsAPI-docs
style: formatting API-docs close parenthesis comment core/settings.py
2023-10-31 13:29:06 +00:00
Varun Dhand
5e0dd5c63b feat: mobile responsive hero page 2023-10-31 18:56:18 +05:30
Alex
f19114e530 Merge pull request #749 from Sai-Suraj-27/fix_codecov
Fixed wrong closing parenthesis in `codecov.yml`
2023-10-31 13:16:15 +00:00
Alex
0db40ecf0f Merge pull request #746 from Sai-Suraj-27/fix_branch_name
Fixed branch name from `master` to `main` in the contributing guide.
2023-10-31 13:10:50 +00:00
Alex
8289067a4e Merge pull request #709 from CBID2/making-colors-accessible
feat: made color accessible
2023-10-31 13:03:54 +00:00
Guspan Tanadi
9c5e3d094b docs: insert Method description /api/docs_check section API-docs 2023-10-31 19:59:09 +07:00
Guspan Tanadi
cb12b19c1e Merge branch 'arc53:main' into settingsAPI-docs 2023-10-31 19:53:59 +07:00
Alex
5d0b8588f9 Revert "delete conflicting checkmark"
This reverts commit 266087c5f1.
2023-10-31 12:52:34 +00:00
Alex
0c05e1036d del 2023-10-31 12:52:27 +00:00
Alex
266087c5f1 delete conflicting checkmark 2023-10-31 12:50:38 +00:00
Alex
147b94d936 checkmark bug 2023-10-31 12:49:25 +00:00
Alex
872511ebb9 Merge pull request #731 from harshita-2003/patch-2
Update README.md
2023-10-31 12:41:55 +00:00
Guspan Tanadi
ce8ed5bfeb style: formatting /api/task_status API-docs 2023-10-31 16:45:23 +07:00
Guspan Tanadi
d81838dfc4 docs: close parentheses EMBEDDINGS_KEY comment settings 2023-10-31 16:40:41 +07:00
Sai-Suraj-27
79ec3594fe Fixed wrong closing parenthesis in codecov.yml 2023-10-31 12:03:13 +05:30
Sai-Suraj-27
cdb246697e Fixed branch name from master to main in contributing guide. 2023-10-31 11:03:04 +05:30
Alex
6476e688e5 Fixes sidebar 2023-10-30 22:41:35 +00:00
Alex
5d1ec6a9c8 Update README.md 2023-10-30 22:07:34 +00:00
Alex
be8a7e981a Update README.md 2023-10-30 22:01:04 +00:00
Alex
d59731a678 Merge pull request #722 from lakshmi930/update-sidebar-ui
Update sidebar UI and Logo
2023-10-30 21:55:00 +00:00
Lakshmi Narayanan
0254510d53 Fix the rotation of the avatar 2023-10-31 00:40:35 +04:00
Lakshmi Narayanan
9327955891 Update sidebar effects and styles based on figma 2023-10-31 00:38:12 +04:00
Christine Belzie
4daf08e20f fix: use a new color 2023-10-30 13:21:53 -04:00
SoumyadiptoPal
6fc31ddedb Updated the code 2023-10-30 22:32:39 +05:30
0xrahul6
fac8c9ee4e Enhancement: Updated Train other docs 2023-10-30 14:53:08 +00:00
0xrahul6
d05f7e2084 Enhanced Guides/Customizi 2023-10-30 14:23:21 +00:00
Alex
0a0a6bae0f Merge pull request #728 from ka1bi4/update/docs-improve-and-fixes
Update/docs improve and fixes
2023-10-30 13:19:42 +00:00
Roman Zhukov
560c063db4 Update Quickstart docs with bash language hl. 2023-10-30 13:20:19 +03:00
Roman Zhukov
54ac2d33e2 Update Quickstart docs with bash language hl. 2023-10-30 13:06:37 +03:00
HARSHITA GUPTA
fb3be8a6a0 Update README.md
added a lighting emoji to give it more great look
2023-10-30 13:00:47 +05:30
Lakshmi Narayanan
5a33953b78 Add Chats heading if there are any conversations 2023-10-30 10:14:40 +04:00
Alex
ba7a8fc796 Merge pull request #710 from mishmanners/main
Make README accessible
2023-10-30 01:12:51 +00:00
Michelle "MishManners®™" Mannering
0bdee8219a Fix README
for some reason - things were missing; fixed now
2023-10-30 11:59:47 +11:00
Alex
f82951f412 Merge pull request #723 from FarukhS52/main
Fix typo
2023-10-30 00:42:02 +00:00
Alex
35e188b851 Merge pull request #719 from theprince29/patch-1
Update CONTRIBUTING.md
2023-10-29 23:38:50 +00:00
Roman Bug
8990e4666a Update Quickstart.md 2023-10-30 01:46:06 +03:00
Roman Bug
ceff618e5d Update API-docs.md 2023-10-30 01:18:52 +03:00
Roman Bug
cf3aab9d38 Update Quickstart.md 2023-10-30 01:06:24 +03:00
Roman Bug
a74c70e8a1 Update README.md 2023-10-30 00:59:17 +03:00
Alex
46817c7664 Merge pull request #679 from akash0708/fix/hero-styling
fix: Hero section styling fixed, made responsive across all devices
2023-10-29 21:40:59 +00:00
Akshay
c0c9cab14c Formatting 2023-10-29 21:38:17 +05:30
Alex
478a034740 Merge pull request #725 from beKoool/remote-upload-ui
Design the Remote sources upload menu UI
2023-10-29 13:57:07 +00:00
beKool.sh
01693cb155 Fix random spaces 2023-10-29 17:13:23 +05:45
Farookh Zaheer Siddiqui
7a44c9e650 Update Railway-Deploying.md 2023-10-29 14:09:56 +05:30
Farookh Zaheer Siddiqui
70a6a275f4 Update CODE_OF_CONDUCT.md 2023-10-29 14:03:05 +05:30
Lakshmi Narayanan
e627ebc127 Small fix with fixed height and width 2023-10-29 02:01:18 +04:00
Lakshmi Narayanan
56b81b78c3 Update sidebar with new logo and icon 2023-10-29 01:48:04 +04:00
Lakshmi Narayanan
c304485079 Fix the color of documentation icon 2023-10-29 01:44:00 +04:00
Lakshmi Narayanan
df51797c29 Add expand icon 2023-10-29 01:43:09 +04:00
Lakshmi Narayanan
754339214c Update logo in conversation 2023-10-29 01:42:26 +04:00
Lakshmi Narayanan
057ecc3ed9 Update logo in Homepage and About page 2023-10-29 01:41:55 +04:00
Lakshmi Narayanan
c14f79ebf7 Add the new logo 2023-10-29 01:40:50 +04:00
Pavel
71fdff17de Merge pull request #721 from arc53/feature/anthropic
anthropic LLM
2023-10-28 22:58:30 +04:00
Alex
04b4001277 anthropic working 2023-10-28 19:51:12 +01:00
PRINCE PAL
fbfb8a3b41 Update CONTRIBUTING.md
Under the workflow heading I found that in point 2 & 3 shell was missing, that I rectified and corrected it
2023-10-28 21:17:25 +05:30
dependabot[bot]
1bee088fe6 Bump urllib3 from 1.26.17 to 1.26.18 in /application
Bumps [urllib3](https://github.com/urllib3/urllib3) from 1.26.17 to 1.26.18.
- [Release notes](https://github.com/urllib3/urllib3/releases)
- [Changelog](https://github.com/urllib3/urllib3/blob/main/CHANGES.rst)
- [Commits](https://github.com/urllib3/urllib3/compare/1.26.17...1.26.18)

---
updated-dependencies:
- dependency-name: urllib3
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-10-28 16:52:37 +02:00
Akash Bag
d2e4d6ecf0 fix: fixed bug due to change in number of lines of text 2023-10-28 11:01:50 +05:30
Akash Bag
5f03f90582 fix: fixed styling for firefox 2023-10-27 16:35:59 +05:30
Alex
e54d46aae1 Update Navigation.tsx 2023-10-27 02:22:04 +01:00
Alex
54a3b9900e fix error msg 2023-10-27 01:52:09 +01:00
Michelle "MishManners®™" Mannering
1dc16e900a Make README accessible
Add accessibility features to README file based on the top tips from GitHub: https://github.blog/2023-10-26-5-tips-for-making-your-github-profile-page-accessible/
2023-10-27 11:22:47 +11:00
Alex
08a7e666b2 Update ConversationTile.tsx 2023-10-27 00:52:15 +01:00
Alex
678fd28f1d Update Railway-Deploying.md 2023-10-26 22:53:57 +01:00
Christine
ff89c3b274 feat: made color accessible 2023-10-26 19:12:33 +00:00
Alex
cff7aebe55 fix 2023-10-26 18:06:26 +01:00
Alex
ed3a3d0876 Merge pull request #706 from debnath003/main
Update some markdown files
2023-10-26 17:53:45 +01:00
Alex
425cd9eb26 Merge pull request #707 from akash0708/docs/typo
docs: Typo in quickstart guide fixed
2023-10-26 17:39:26 +01:00
Alex
ebe84dd8a4 Merge pull request #703 from Ankit-Matth/revamp_icon_effects
Background color of like/dislike has been changed
2023-10-26 10:39:52 +01:00
Akash Bag
217c4144b5 docs: Typo in quickstart guide fixed 2023-10-26 13:09:08 +05:30
Pronay Debnath
aaeed64621 Update and rename Hosting-DocsGPT.md to Hosting-the-app.md 2023-10-26 11:17:19 +05:30
Pronay Debnath
9133a56d2a Update _meta.json 2023-10-26 11:16:26 +05:30
Pronay Debnath
12bd7dc44f Update _meta.json 2023-10-26 11:14:30 +05:30
Pronay Debnath
32fa86adaa Update _meta.json 2023-10-26 11:11:41 +05:30
Pronay Debnath
1811cff1f9 Rename Hosting-the-app.md to Hosting-DocsGPT.md 2023-10-26 10:59:32 +05:30
Pronay Debnath
2e6a5c0525 Update My-AI-answers-questions-using-external-knowledge.md 2023-10-26 10:57:24 +05:30
Pronay Debnath
b68b214d08 Update My-AI-answers-questions-using-external-knowledge.md 2023-10-26 10:56:57 +05:30
Pronay Debnath
153d12d93f Update How-to-use-different-LLM.md 2023-10-26 10:47:35 +05:30
Pronay Debnath
1a62a773ae Update react-widget.md 2023-10-26 10:40:19 +05:30
Pronay Debnath
5f0cccb81e Update Customising-prompts.md 2023-10-26 10:37:58 +05:30
Ankit Matth
5c7f4b3df7 Merge branch 'arc53:main' into revamp_icon_effects 2023-10-26 07:17:40 +05:30
Ankit Matth
1248b76b41 Background color of like/dislike has been changed 2023-10-26 06:20:50 +05:30
Alex
c4176af1ea Merge pull request #672 from Ankit-Matth/revamp_icon_effects
[feature] : The UI of the feedback icons has been changed.
2023-10-25 22:29:50 +01:00
Alex
799c306138 Merge branch 'main' of https://github.com/arc53/DocsGPT 2023-10-25 22:27:12 +01:00
Alex
c65e3fdf62 Update _meta.json 2023-10-25 22:27:11 +01:00
Alex
08712ef4f8 Merge pull request #688 from SoumyadiptoPal/stickyNavbar
Feature: left side toggle button should be sticky
2023-10-25 22:23:57 +01:00
Alex
d0119f5bf1 Merge pull request #694 from Ram-tyagi/main
[Add] Vite and react documentation link
2023-10-25 22:05:59 +01:00
Alex
f3c626c800 Merge pull request #556 from Exterminator11/openapi3_parser
Parser for OpenAPI3(Swagger)
2023-10-25 21:27:45 +01:00
Alex
f1891478d5 Merge pull request #692 from Yash-sudo-web/main
Enhanced How-to-use-different-LLM.md
2023-10-25 21:26:32 +01:00
Ankit Matth
7a2e6e640d [making required changes] - Update ConversationBubble.tsx 2023-10-25 21:48:25 +05:30
Soumyadipto Pal
890d418639 Merge branch 'main' into stickyNavbar 2023-10-25 21:07:53 +05:30
Ram-tyagi
f38c934a6d [Add] Vite and react documentation link 2023-10-25 17:42:01 +05:30
Exterminator11
f3540aac0f Changed import 2023-10-25 17:07:47 +05:30
Exterminator11
889ce984a9 Made changes 2023-10-25 16:50:01 +05:30
Alex
89a437149c Merge pull request #690 from thanmaisai/main
Update README.md
2023-10-25 07:19:04 -04:00
Alex
43f65651ac Merge pull request #675 from krishvsoni/patch-2
Patch 2
2023-10-25 06:57:29 -04:00
Alex
d74d69c1c8 Merge pull request #681 from nobunagaaa/docs/enhance
Enhancement: Update react-widget.md docs
2023-10-25 06:52:53 -04:00
Alex
fefc85683c Merge pull request #677 from AnkitaSikdar005/code4
Fixed Typos and minor issues in CODE_OF_CONDUCT.md
2023-10-25 06:41:37 -04:00
Alex
6f97158c0e Merge pull request #676 from Gourav2609/main
Fixed overflow
2023-10-25 06:39:30 -04:00
Alex
1320101112 Merge pull request #663 from neha3423/main
Updated workflow of Contributing.md
2023-10-25 06:32:11 -04:00
Alex
031a267394 Merge pull request #537 from AnkitaSikdar005/code1
Updated CONTRIBUTING.md
2023-10-25 06:30:41 -04:00
Alex
9119030959 Merge pull request #670 from iamakhileshmishra/main
Railway Deployment Guide Added
2023-10-25 06:26:09 -04:00
Yash-sudo-web
9b10a8028d Enhanced How-to-use-different-LLM.md 2023-10-25 15:38:19 +05:30
SoumyadiptoPal
4be38fcb0e Settings page added 2023-10-25 10:24:55 +05:30
noobcoder
9090f4485a Update README.md 2023-10-24 19:31:39 +00:00
iamakhileshmishra
5749d66ac9 Merge branch 'main' of https://github.com/iamakhileshmishra/DocsGPT 2023-10-24 23:59:42 +05:30
Akhilesh Kumar Mishra
4bb4b4eb1d Merge branch 'arc53:main' into main 2023-10-24 23:59:31 +05:30
iamakhileshmishra
103d062f74 package conflict solved 2023-10-24 23:57:49 +05:30
SoumyadiptoPal
9893480089 Made the navbar sticky 2023-10-24 22:27:55 +05:30
Alex
5dbd240017 Update DocsGPT tee-back.jpeg 2023-10-24 10:51:28 +01:00
nobunagaaa
e0dce8fd01 Enhancement: Update react-widget.md docs 2023-10-24 15:05:43 +05:30
Akash Bag
492139942c fix: Hero section styling fixed, made responsive accross all devices 2023-10-24 10:38:51 +05:30
Exterminator11
8ebff1a908 Updated test_openapi3parser.py 2023-10-24 07:43:57 +05:30
Ankita Sikdar
44def1f6bc Update CODE_OF_CONDUCT.md 2023-10-24 01:05:53 +05:30
Gourav2609
8934b9ab5c Fixed overflow 2023-10-23 23:15:07 +05:30
KRISH SONI
130a6b67bd Update README.md 2023-10-23 22:16:19 +05:30
Alex
2df32cd9a7 Merge pull request #668 from lakshmi930/fix-small-ui-bugs
Fix something went wrong message bubble
2023-10-23 12:43:22 -04:00
Alex
d413d58b47 Merge pull request #652 from rasvanjaya21/main
Enhance backdrop modal effect
2023-10-23 12:38:06 -04:00
KRISH SONI
9e632aa0bd Update README.md 2023-10-23 22:01:07 +05:30
Alex
964020ee12 Merge pull request #651 from 0xrahul6/docs/enhance
Enhanced Chatwoot-Extension
2023-10-23 12:10:16 -04:00
Alex
672e14d6ea Merge pull request #641 from Ankit-Matth/documentation_icon_changed
I have changed icon for Documentation in the left side bar.
2023-10-23 11:59:34 -04:00
Alex
54baf04a86 Merge pull request #658 from RDxR10/main
Update react-widget.md
2023-10-23 11:56:03 -04:00
Alex
64c83460b9 Merge pull request #656 from Lokendrakushwah12/main
added the 3 cards in hero section with gradient border also its respo…
2023-10-23 11:52:49 -04:00
Alex
74ec3fa7d4 Merge pull request #655 from parthrc/adding-custom-404-page
Added a custom 404 not found page
2023-10-23 11:44:24 -04:00
Alex
0821d7a803 Merge pull request #654 from adarsh-jha-dev/patch-2
Enhanced Hooks with Outside Click Handling and Dark Mode Detection
2023-10-23 11:40:59 -04:00
Ankita Sikdar
55beb3978c Merge branch 'main' into code1 2023-10-23 17:30:18 +05:30
Ankita Sikdar
9b044815de Update CODE_OF_CONDUCT.md 2023-10-23 17:07:25 +05:30
Ankit Matth
0668fea3b7 Revamp Icon Effects 2023-10-23 17:01:41 +05:30
iamakhileshmishra
a6677b2e45 Railway Deployment Guide Added 2023-10-23 14:44:37 +05:30
Yusril A
6f544f56d8 Merge branch 'arc53:main' into main 2023-10-23 11:20:04 +07:00
Lakshmi Narayanan
5556be9cab Fix something went wrong message bubble 2023-10-23 00:52:01 +04:00
Alex
465c4afe8d Merge pull request #653 from Ankit-Matth/aboutPage_alignment
I have aligned about page in the center.
2023-10-22 11:59:39 -04:00
Alex
78dd1e1d81 Merge branch 'main' into aboutPage_alignment 2023-10-22 11:56:13 -04:00
Alex
eebfc78ad3 Merge pull request #626 from shruti-sen2004/edit_2
Update react-widget.md
2023-10-22 11:53:04 -04:00
Alex
4783685fdb Merge pull request #650 from SamsShow/main
[Bug]:About section not Aligned.
2023-10-22 11:51:13 -04:00
Alex
bfd0363fad Merge pull request #647 from beKoool/settings-ui
Design the Settings UI
2023-10-22 11:48:21 -04:00
KRISH SONI
e9323ba2ec Merge branch 'main' into patch-2 2023-10-22 21:17:56 +05:30
Alex
dac774c9d2 Merge pull request #639 from Ayush-Prabhu/patch-2
Update How-to-use-different-LLM.md
2023-10-22 11:39:44 -04:00
Alex
664ee2b433 Merge pull request #637 from debghs/debghs-patch-1
Update How-to-train-on-other-documentation.md
2023-10-22 11:36:50 -04:00
Alex
85f283fe2b Merge pull request #636 from Ritish134/patch-1
Update README.md
2023-10-22 11:33:43 -04:00
Alex
7bd7d66afc Merge branch 'main' into patch-1 2023-10-22 11:31:51 -04:00
Alex
81b16aa900 Merge pull request #638 from Bitnagar/fix(frontend)--fix-header-z-index
Fix(frontend):  fix Navigation component z-index for small devices
2023-10-22 11:27:25 -04:00
Alex
e07fb34ace Merge pull request #504 from ayan-joshi/New-prompt-
Improving Customising_prompt.md file
2023-10-22 11:19:03 -04:00
Alex
9303746d80 Merge branch 'main' into New-prompt- 2023-10-22 11:16:03 -04:00
Alex
e3e8e67cb4 Merge pull request #627 from YadlaMani/main
Fixed Typo
2023-10-22 11:15:00 -04:00
Alex
6490027e57 Merge pull request #577 from shivanandmn/faiss_delete_index
added delete index of vector store in faiss
2023-10-22 10:15:31 -04:00
neha3423
6cbe4f2ea7 neha3423 2023-10-22 11:38:23 +05:30
neha3423
960365a063 neha3423 2023-10-22 10:54:02 +05:30
RDxR10
839d614c9c Update react-widget.md 2023-10-21 12:53:47 +05:30
Lokendra Kushwah
ae13e557a7 added the 3 cards in hero section with gradient border also its responsive 2023-10-20 23:24:56 +05:30
unknown
a245383f8c Added a custom 404 not found page 2023-10-20 16:22:35 +05:30
Adarsh Jha
78b8d3e41d Update index.ts 2023-10-20 14:33:43 +05:30
Ankit Matth
fcfaa04cc6 About page aligned in center 2023-10-20 05:42:12 +05:30
rasvanjaya21
fe866b2d66 Enhance backdrop modal effect 2023-10-20 02:57:41 +07:00
0xrahul6
e7bbc4ac0c Enhanced Chatwoot-Extension 2023-10-19 17:13:30 +00:00
Saksham Tyagi
3b746c91df Merge branch 'main' of https://github.com/SamsShow/DocsGPT 2023-10-19 20:15:45 +05:30
Saksham Tyagi
06f0129b59 💄 About margin 2023-10-19 20:12:40 +05:30
Alex
641e75b8a8 Merge pull request #620 from rahul0x00/docs/API-docs
Enhancement: Improve API Endpoint Documentation
2023-10-19 07:41:17 -04:00
Alex
35f9fda457 Merge pull request #633 from Rutam21/patch-1
Added Civo Cloud Deployment option.
2023-10-19 07:38:42 -04:00
Alex
de29d69efe Merge pull request #619 from vedant-z/patch-1
fix: Direction of dropdown arrow corrected
2023-10-19 07:37:19 -04:00
Alex
f587af1005 Merge pull request #625 from krabbi/prevent_unnecessary_renders
[FE]Prevent unneccessary renders.
2023-10-19 07:30:56 -04:00
Alex
4ed6580e1d Merge pull request #623 from krabbi/fix_lists_in_converastion
[FE]: Fix render lists in conversation
2023-10-19 07:29:02 -04:00
beKool.sh
2f6213c944 Change font weight 2023-10-19 11:11:30 +05:45
KRISH SONI
f365b76cfc Merge branch 'main' into patch-2 2023-10-18 21:58:17 +05:30
KRISH SONI
55921b262f docker docs 2023-10-18 21:54:28 +05:30
Alex
3039c97989 Merge pull request #622 from Yash-sudo-web/main
Updated Chatwoot-extension.md
2023-10-18 11:30:47 -04:00
Alex
a1af4f19c5 Merge pull request #621 from arc53/dependabot/npm_and_yarn/frontend/babel/traverse-7.23.2
Bump @babel/traverse from 7.20.13 to 7.23.2 in /frontend
2023-10-18 11:13:48 -04:00
Alex
131e4087fd Merge pull request #631 from Raunakkumarr/docsCorrection-selfHosting
Correction in docs
2023-10-18 11:11:11 -04:00
Ankit Matth
ee6471351d Icon for Documentation changed 2023-10-18 18:41:03 +05:30
Ayush-Prabhu
d93266fee2 Update How-to-use-different-LLM.md
Corrected grammatical errors to remove ambiguity and improve professionalism.
2023-10-18 16:21:15 +05:30
Shivam Bhatnagar
dbbf39db6d pre-commit hook 2023-10-18 10:34:50 +00:00
Shivam Bhatnagar
d40ea44ae6 fix(frontend): fix navigation z-index in mobiles 2023-10-18 10:33:44 +00:00
debghs
f0d4847946 Update How-to-train-on-other-documentation.md 2023-10-18 12:46:29 +05:30
Ritish Srivastava
98a9c766ef Update README.md
Enhance grammar and formatting in various sections.
2023-10-18 07:49:24 +05:30
Rutam Prita Mishra
91393b650b Added Civo Cloud Deployment option.
This change adds a new deployment guide for Civo Compute Cloud.
2023-10-18 04:09:50 +05:30
Alex
49a4b119e1 Merge pull request #615 from shelar1423/patch3
FIX: added script in package.json of docs and added instructions to run nextra-DocsGPT locally
2023-10-17 14:21:51 -04:00
Alex
e69fab822b Merge pull request #624 from krabbi/fix_footer_width
[FE]Fix footer width ob md+ screens
2023-10-17 14:19:31 -04:00
Raunak Kumar
45c58cc766 Correction in docs 2023-10-17 23:31:13 +05:45
ManiYadla
ca48f000bd changed the typos 2023-10-17 20:40:27 +05:30
Shivanand
21ba1e3958 Merge branch 'main' into faiss_delete_index 2023-10-17 17:35:30 +05:30
Rahul Kumar
062f3256a7 Update API-docs.md 2023-10-17 12:28:50 +05:30
Rahul Kumar
186f565b99 Update API-docs.md 2023-10-17 12:27:38 +05:30
Vedant Borkar
5c2b4398d9 Update Navigation.tsx 2023-10-17 11:53:57 +05:30
Rahul Kumar
a9fb61bbd6 Update API-docs.md 2023-10-17 08:28:13 +05:30
Shruti Sen
a51e25dbde Update react-widget.md 2023-10-17 08:09:32 +05:30
Alexander Deshkevich
0a717ae82e Prevent unneccessary renders. Update show/hide state of feedback buttons by css instead React 2023-10-16 20:13:20 -03:00
Alexander Deshkevich
f9e6751279 fix render lists in conversation 2023-10-16 18:59:31 -03:00
Alexander Deshkevich
0306f8ec65 Fix footer width ob md+ screens 2023-10-16 16:35:33 -03:00
Alex
66f2e549ce Shirt preview 2023-10-16 15:02:28 -04:00
Alex
9ab413643a Update HACKTOBERFEST.md 2023-10-16 15:02:02 -04:00
Alex
3a4eeb01b0 Update CONTRIBUTING.md 2023-10-16 15:00:34 -04:00
Alex
57a8dcc155 Merge pull request #616 from Sanyam-2026/main
Fixed Typo
2023-10-16 13:59:07 -05:00
Yash-sudo-web
2f21476b2a Update Chatwoot-extension.md 2023-10-17 00:18:40 +05:30
Yash-sudo-web
9f9e2f3b24 Update Chatwoot-extension.md 2023-10-17 00:15:03 +05:30
dhselar1423
f886dfb60c removed commandline 2023-10-16 23:28:39 +05:30
dependabot[bot]
c22b014056 Bump @babel/traverse from 7.20.13 to 7.23.2 in /frontend
Bumps [@babel/traverse](https://github.com/babel/babel/tree/HEAD/packages/babel-traverse) from 7.20.13 to 7.23.2.
- [Release notes](https://github.com/babel/babel/releases)
- [Changelog](https://github.com/babel/babel/blob/main/CHANGELOG.md)
- [Commits](https://github.com/babel/babel/commits/v7.23.2/packages/babel-traverse)

---
updated-dependencies:
- dependency-name: "@babel/traverse"
  dependency-type: indirect
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-10-16 17:53:36 +00:00
Alex
d899b6a7e1 Merge pull request #588 from asoderlind/fix/as/embedding-size-mismatch
raise more legible error if the word embedding dimensions don't match
2023-10-16 12:53:08 -05:00
Alex
450dde3739 Merge pull request #614 from 5h0ov/patch-3
Update Chatwoot-extension.md
2023-10-16 12:43:05 -05:00
Alex
2ac40903f3 Update README.md 2023-10-16 13:36:54 -04:00
Exterminator11
f328b39f57 Fixed import error 2023-10-16 21:05:47 +05:30
Rahul Kumar
06cc4b07ab Update API-docs.md 2023-10-16 20:35:17 +05:30
Rahul Kumar
1c0b68f0e3 Update API-docs.md 2023-10-16 20:27:09 +05:30
Rahul Kumar
efcce6a826 Enhancement: Improve API Endpoint Documentation 2023-10-16 14:51:24 +00:00
Vedant Borkar
fa8177d0e5 Update Navigation.tsx 2023-10-16 19:52:16 +05:30
asoderlind
d51cd8df89 add file docstring 2023-10-16 11:31:10 +02:00
Sanyam Jain
5530d611b9 Update Hosting-the-app.md 2023-10-16 15:00:11 +05:30
asoderlind
e73636bef3 remove trailing whitespace, sort imports, remove unused arguments 2023-10-16 11:22:42 +02:00
dhselar1423
74ff994281 nextra-docsGPT local setup 2023-10-16 10:59:25 +05:30
dhselar1423
7b28d353ee made changes in package.json and added instructions nextra-docsgpt 2023-10-16 10:57:49 +05:30
asoderlind
e2a8ca143a remove unused imports 2023-10-16 06:13:26 +02:00
asoderlind
4e81f98927 add dependency 2023-10-16 06:13:15 +02:00
Shuvadipta Das
ab4c994266 Update Chatwoot-extension.md 2023-10-16 08:56:38 +05:30
Alex
aea6a434f1 Merge pull request #610 from rahul0x00/enhance/quickstart
Quickstart: Enhanced Documentation for Launching DocsGPT and Installing Chrome Extension
2023-10-15 16:48:08 -05:00
Alex
6c95d5a2de Merge pull request #579 from ankur0904/contribution-guideline
Adds contributing guidelines with steps
2023-10-15 16:46:19 -05:00
Alex
fcaabb2c1e Merge pull request #595 from HeetVekariya/main
fix: Remove extra spaces in response
2023-10-15 16:44:40 -05:00
Alex
66b2722cad Merge pull request #609 from shelar1423/patch2
CONTRIBUTING.MD : Added DocsGPT -Hacktoberfest t-shirt images
2023-10-15 16:37:00 -05:00
shivanandmn
2e95666939 added new endpoint 2023-10-16 02:46:48 +05:30
Akshay
cdfcd99695 Fix docs_check and upload endpoint 2023-10-15 21:31:02 +05:30
Akshay
e71d21fc27 Fix render method 2023-10-15 20:23:07 +05:30
Akshay
e95ebfd6a0 Add docker files 2023-10-15 20:23:07 +05:30
Akshay
e5a875856a Minor 2023-10-15 20:23:06 +05:30
Akshay
930218c067 Mock task status call 2023-10-15 20:23:06 +05:30
Akshay
ff1362e462 Mock upload call 2023-10-15 20:23:06 +05:30
Akshay
01457bbe79 Add feedback endpoint 2023-10-15 20:23:06 +05:30
Akshay
8c7da0bdb6 Add more routes 2023-10-15 20:23:06 +05:30
Akshay
6f634c3f13 Add mock server module 2023-10-15 20:23:06 +05:30
Exterminator11
a7f5303eaf Cleaned up the code 2023-10-15 17:20:50 +05:30
Exterminator11
7159e4fbe2 Formatted files 2023-10-15 17:16:58 +05:30
Exterminator11
36b243e9d2 Formatted all the changed files 2023-10-15 17:16:12 +05:30
Exterminator11
bd70e00f08 Added tests and updated openapi3_parser.py 2023-10-15 17:00:54 +05:30
asoderlind
0ca96130c8 remove trailing whitespace 2023-10-15 10:23:09 +02:00
asoderlind
09aa56b63d add test 2023-10-15 10:22:07 +02:00
asoderlind
60cd6a455a refactor 2023-10-15 10:22:00 +02:00
asoderlind
4752ce5250 fix linting error 2023-10-15 09:12:00 +02:00
Rahul Kumar
832569a79c Update Quickstart.md 2023-10-15 09:54:25 +05:30
Rahul Kumar
ecd8cebbef Quickstart: Enhanced Documentation for Launching DocsGPT and Installing Chrome Extension 2023-10-15 04:20:09 +00:00
ankur0904
3c37efa650 Adds the Note for testing 2023-10-15 09:41:27 +05:30
Ankur Singh
21b6ce204d Merge branch 'arc53:main' into contribution-guideline 2023-10-15 09:32:43 +05:30
HeetVekariya
337d2970a0 fix: removed px-2 for source 2023-10-15 09:03:28 +05:30
HeetVekariya
3e5bd25c6e fix: removed items-center from conversationBubble.tsx 2023-10-15 08:47:55 +05:30
Digvijay Shelar
7f0f68b707 Delete Assets/images.md 2023-10-15 02:35:49 +05:30
Digvijay Shelar
ea85482736 Update CONTRIBUTING.md 2023-10-15 02:35:23 +05:30
Digvijay Shelar
01160a5361 Add files via upload 2023-10-15 02:31:45 +05:30
Digvijay Shelar
f4b5a02197 Create images.md 2023-10-15 02:31:08 +05:30
Alex
f724f10a35 Merge pull request #607 from mozi47/fix/sidebar-glitch
sidebar glitch fixed
2023-10-14 15:05:48 -05:00
Alex
0c221ba3d7 Merge pull request #554 from arc53/dependabot/pip/scripts/langchain-0.0.312
Bump langchain from 0.0.308 to 0.0.312 in /scripts
2023-10-14 15:03:38 -05:00
Alex
1907aaf32f Merge pull request #580 from DHRUVKADAM22/DHRUVKADAM22-patch-5
Dhruvkadam22 patch 5
2023-10-14 14:59:55 -05:00
Alex
d6f26b3133 Merge pull request #574 from rahul0x00/enhance/deployment_docs
Enhanced Documentation for Self-Hosting DocsGPT on Amazon Lightsail
2023-10-14 14:53:39 -05:00
Alex
c97a55e65f Merge pull request #460 from alienishi/main
Modified README resolving issue #457
2023-10-14 14:52:08 -05:00
Alex
4a6e38f7da Merge pull request #602 from siddwarr/patch-1
Update README.md
2023-10-14 14:50:48 -05:00
Alex
845ef42338 Merge pull request #598 from Juneezee/simplify-jsx-conditional-rendering
refactor(frontend): simplify JSX conditional rendering
2023-10-14 14:48:29 -05:00
Alex
fde8de8b9e Merge pull request #558 from MSaiKiran9/train-button
Train button Disabled Before Selecting File
2023-10-14 14:43:50 -05:00
Muzakir Shah
88123261ac sidebar glitch fixed 2023-10-15 00:38:46 +05:00
Alex
c04b76528b Merge pull request #597 from shelar1423/main
[Docs] :Enhance Tech Stack Overview with Emojis and Bullet Points
2023-10-14 14:28:18 -05:00
dependabot[bot]
04a13c2ebb Bump langchain from 0.0.308 to 0.0.312 in /scripts
Bumps [langchain](https://github.com/langchain-ai/langchain) from 0.0.308 to 0.0.312.
- [Release notes](https://github.com/langchain-ai/langchain/releases)
- [Commits](https://github.com/langchain-ai/langchain/compare/v0.0.308...v0.0.312)

---
updated-dependencies:
- dependency-name: langchain
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-10-14 19:20:49 +00:00
Alex
6b3cc62cbe Merge pull request #553 from arc53/dependabot/pip/application/langchain-0.0.312
Bump langchain from 0.0.308 to 0.0.312 in /application
2023-10-14 14:20:07 -05:00
Alex
8627be07e7 Merge pull request #592 from faria-karim-porna/ui-sidebar-chat-options
Fix the font size and alignment of all the chat options of the sidebar based on issue #586
2023-10-14 14:13:58 -05:00
Alex
5509a5bca3 Merge pull request #589 from outlivo/outlivo-UI
Closing #357 using CONTRIBUTING.md
2023-10-14 14:10:47 -05:00
Alex
dd52949a2a Merge pull request #510 from HarshMN2345/patch-2
Update API-docs.md
2023-10-14 13:53:53 -05:00
Alex
a310ae6566 Merge branch 'main' into patch-2 2023-10-14 13:52:25 -05:00
Aditya Aryaman Das
1f8643c538 Updated README.md 2023-10-14 19:03:09 +05:30
Aditya Aryaman Das
6ea313970d Merge branch 'arc53:main' into main 2023-10-14 19:01:02 +05:30
Siddharth Warrier
13e6b15308 Update README.md 2023-10-14 14:53:38 +05:30
Eng Zer Jun
0efc2277dd refactor(frontend): simplify JSX conditional rendering
JSX conditional rendering can be simplified to use the logical AND
operator (&&) [1] instead of ternary operator (? :) if we want to render
something only when the conditon is true, and nothing otherwise.

[1]: https://react.dev/learn/conditional-rendering#logical-and-operator-

Signed-off-by: Eng Zer Jun <engzerjun@gmail.com>
2023-10-14 13:39:23 +08:00
Digvijay Shelar
9e6aecd707 Update CONTRIBUTING.md 2023-10-13 18:35:13 +05:30
Digvijay Shelar
f5510ef1b5 Update CONTRIBUTING.md 2023-10-13 18:32:56 +05:30
HeetVekariya
c8e6224946 fix: else text corrected 2023-10-13 09:46:07 +05:30
HeetVekariya
bf11300ab3 fix: Remove extra spaces in response 2023-10-13 09:33:55 +05:30
faria-karim-porna
7361a35c94 Fix the font size and alignment of all the chat options of the sidebar based on issue #586
Fix the font size and alignment of all the chat options of the sidebar based on issue #586
2023-10-13 06:01:27 +06:00
Outlivo
02b2cebb85 Solving issue #357 using CONTRIBUTING.md 2023-10-13 01:07:49 +05:30
asoderlind
9b6ae46e92 raise more legible error if the embedding vector dims don't match 2023-10-12 20:24:25 +02:00
Alex
e5e5a42736 Update CONTRIBUTING.md 2023-10-12 18:19:40 +01:00
Alex
308d8afe4e Update CONTRIBUTING.md 2023-10-12 18:19:06 +01:00
Rahul Kumar
2100cd77ce Update Hosting-the-app.md 2023-10-12 18:52:54 +05:30
Harsh Mahajan
58b13ae69a Update API-docs.md 2023-10-12 17:31:41 +05:30
M Sai Kiran
993c9b31bd Update Upload.tsx 2023-10-12 15:10:26 +05:30
M Sai Kiran
b5d6f0ad36 Update Upload.tsx 2023-10-12 15:07:48 +05:30
M Sai Kiran
03c05a82e4 Merge branch 'arc53:main' into train-button 2023-10-12 14:59:08 +05:30
DHRUVKADAM22
cc887d25e4 Update Quickstart.md
added windows Quickstart guide in Quickstart.md for easy to understand and make more user friendly
2023-10-12 14:48:44 +05:30
ankur0904
80e2d0651b Adds contributing guidelines with steps 2023-10-12 14:32:44 +05:30
shivanandmn
ca3e549dd4 added delete index of vector store in faiss 2023-10-12 13:29:52 +05:30
Rahul Kumar
51f2ca72b9 Enhanced Documentation for Self-Hosting DocsGPT on Amazon Lightsail 2023-10-12 03:19:40 +00:00
Alex
771950f1de Merge pull request #573 from arc53/bug/update-and-delete
fix update and delete bug
2023-10-11 23:16:44 +01:00
Alex
c969e9c014 fix update and delete bug 2023-10-11 23:11:08 +01:00
Alex
344692f9f6 Merge pull request #542 from shruti-sen2004/main
Update CONTRIBUTING.md
2023-10-11 18:57:59 +01:00
DHRUVKADAM22
fd083078fc Merge pull request #2 from DHRUVKADAM22/DHRUVKADAM22-patch-3
Update README.md
2023-10-11 15:07:00 +05:30
DHRUVKADAM22
9bacae4b2e Update README.md 2023-10-11 15:06:44 +05:30
M Sai Kiran
76baa6c5f8 Before File Selection Train Disabled 1 2023-10-11 09:39:04 +05:30
M Sai Kiran
84c822a0ca Before File Selection Train Button Disabled 2023-10-11 09:38:15 +05:30
Exterminator11
ddd938fd64 Parser for OpenAPI3(Swagger) 2023-10-11 07:36:37 +05:30
Alex
e91b30f4c7 Merge pull request #549 from hariraghav10/ui-update
UI update - Added autofocus for user chat input
2023-10-10 22:31:46 +01:00
dependabot[bot]
31fb1801d2 Bump langchain from 0.0.308 to 0.0.312 in /application
Bumps [langchain](https://github.com/langchain-ai/langchain) from 0.0.308 to 0.0.312.
- [Release notes](https://github.com/langchain-ai/langchain/releases)
- [Commits](https://github.com/langchain-ai/langchain/compare/v0.0.308...v0.0.312)

---
updated-dependencies:
- dependency-name: langchain
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-10-10 21:18:11 +00:00
DHRUVKADAM22
117d0f2e38 Merge pull request #1 from DHRUVKADAM22/DHRUVKADAM22-patch-1
Update README.md
2023-10-10 23:32:10 +05:30
DHRUVKADAM22
79bb79debc Update README.md
I have added quick start guide for windows more specifically 
makes more user friendly by adding basic info and description
(which make user or newbies to get understand properly.)
2023-10-10 23:29:46 +05:30
Harsh Mahajan
11cd022965 Update API-docs.md 2023-10-10 23:22:11 +05:30
Ayan Joshi
d7b28a3586 Update Customising-prompts.md 2023-10-10 19:10:24 +05:30
Shruti Sen
dc14245105 Update CONTRIBUTING.md 2023-10-10 18:05:19 +05:30
Hari Raghav
e772dfaa12 Merge branch 'arc53:main' into ui-update 2023-10-10 17:53:48 +05:30
Shruti Sen
4d29cae936 Update CONTRIBUTING.md 2023-10-10 17:39:07 +05:30
hariraghav10
71ed0ffe13 Merge remote-tracking branch 'origin/ui-update' into ui-update 2023-10-10 17:35:35 +05:30
Shruti Sen
56d0981cee Update CONTRIBUTING.md 2023-10-10 17:35:18 +05:30
Shruti Sen
ad43d10ce4 Merge branch 'arc53:main' into main 2023-10-10 17:34:41 +05:30
Alex
fb6618181a Merge pull request #538 from ManishMadan2882/main
UI bug fixed - Placed the feedback/copy buttons at the top
2023-10-10 12:57:04 +01:00
Shruti Sen
43a9bc0d7b Update CONTRIBUTING.md 2023-10-10 17:22:38 +05:30
hariraghav10
f835b14902 added autofocus for user chat input 2023-10-10 17:22:10 +05:30
Shruti Sen
c1c591d1eb Merge branch 'arc53:main' into main 2023-10-10 17:21:19 +05:30
Hari Raghav
4348549f2d Merge branch 'arc53:main' into ui-update 2023-10-10 17:20:07 +05:30
hariraghav10
e48df87e06 added autofocus feature for chat input div 2023-10-10 17:04:05 +05:30
Alex
e718feb1f7 Update README.md 2023-10-10 11:36:54 +01:00
Shruti Sen
3b6f3f13d4 Update CONTRIBUTING.md 2023-10-10 12:12:25 +05:30
ManishMadan2882
13fabaf6aa bugs fixed 2023-10-10 06:01:21 +05:30
Alex
9cfcdb1c23 Merge pull request #535 from vn-os/main
Correct instructions in `Development environments`
2023-10-09 23:59:14 +01:00
Alex
2800d0dcd3 Merge pull request #532 from staticGuru/error-state-#484
Fix the error state UI
2023-10-09 23:54:14 +01:00
Alex
3e2055255e Merge branch 'main' into error-state-#484 2023-10-09 23:53:14 +01:00
Alex
64a8857884 Merge pull request #529 from staticGuru/chatroom-rename-#495
Chatroom rename feature
2023-10-09 23:42:45 +01:00
Alex
808b291c2c Merge pull request #530 from robbiebusinessacc/patch-1
Backend Code Refactoring for Better Readability and Scalability
2023-10-09 23:40:10 +01:00
Alex
cae3e7136e Merge pull request #525 from Vaibhav91one/bug/text-out-of-screen
bug(front): The text seems to go out of screen when the result is too big.
2023-10-09 23:32:34 +01:00
Alex
c069a187f8 Merge pull request #476 from mnickrogers/main
Fix missing documentation for using Llama_cpp
2023-10-09 23:26:38 +01:00
Ankita Sikdar
91fa932168 Update CONTRIBUTING.md 2023-10-10 02:06:52 +05:30
Vic P
188158a29b Merge branch 'arc53:main' into main 2023-10-10 00:47:26 +07:00
Vic P
a3d5cb5851 Update README.md to correct instructions in Development environments. 2023-10-10 00:47:15 +07:00
Alex
0788582528 Merge pull request #522 from 5h0ov/patch-2
Updated My-AI-answers-questions-using-external-knowledge.md
2023-10-09 16:40:59 +01:00
Alex
da81abc12e Merge pull request #437 from shelar1423/main
[Enhanced]: Useful Links Section in README.md
2023-10-09 16:34:25 +01:00
Alex
81b92111ca Merge pull request #511 from beKoool/docs-fix
[API Docs] Add Backlink to "Vector Stores" and fix grammatical error
2023-10-09 16:32:50 +01:00
Alex
a809e72704 Merge branch 'main' into docs-fix 2023-10-09 16:30:11 +01:00
Alex
cb0e4b6e87 Update README.md 2023-10-09 15:37:40 +01:00
staticGuru
16df8d803c Fix the error state issues 2023-10-09 19:48:35 +05:30
staticGuru
ce7ac78b42 add the darkmode in config file 2023-10-09 19:48:17 +05:30
Harsh Mahajan
c21e0755b3 Update API-docs.md 2023-10-09 17:55:41 +05:30
Harsh Mahajan
e1dc0a576d Update API-docs.md 2023-10-09 17:45:50 +05:30
staticGuru
a998db0570 add fetch conversations in the delete callbacks 2023-10-09 16:30:09 +05:30
staticGuru
c79ec45adb Fix the lint issues 2023-10-09 16:16:56 +05:30
staticGuru
72481e8453 Fix the post API issues 2023-10-09 16:16:20 +05:30
staticGuru
3753f7d138 change the input outline border color 2023-10-09 16:12:56 +05:30
Robbie Walmsley
4d92606562 Update ingest.py 2023-10-09 11:11:07 +01:00
Robbie Walmsley
2d0b6bcfcc Update ingest.py 2023-10-09 11:04:55 +01:00
Robbie Walmsley
57fb29b600 Update worker.py 2023-10-09 10:55:34 +01:00
Robbie Walmsley
340647cb22 Update ingest.py 2023-10-09 10:53:03 +01:00
staticGuru
a06369dd7b add the checkmark icons 2023-10-09 15:19:29 +05:30
staticGuru
95fe103718 add the conversation in the result response 2023-10-09 15:17:32 +05:30
staticGuru
036297ef36 Merge branch 'main' of https://github.com/staticGuru/DocsGPT into chatroom-rename-#495 2023-10-09 14:58:21 +05:30
staticGuru
129c055fee Change the width in the tile 2023-10-09 14:44:43 +05:30
Ayan Joshi
c688656607 Commit 2023-10-09 14:27:59 +05:30
staticGuru
b49e8deb3e add the typescript props interface 2023-10-09 14:13:09 +05:30
staticGuru
17264e7872 add the outside click listioner 2023-10-09 12:50:12 +05:30
staticGuru
022c0c3a89 add trash icons changes 2023-10-09 12:49:09 +05:30
staticGuru
b8539122ed add the update conversation callbacks 2023-10-09 12:48:48 +05:30
Vaibhav91one
4ca906a518 bug(front): The text seems to go out of screen when the result is too big, needed some tailwind 2023-10-09 10:44:14 +05:30
Nick Rogers
7bf67869b0 Reference DocsGPT model in custom model steps. 2023-10-08 22:03:32 -07:00
staticGuru
a032164a99 Add update conversation name API 2023-10-09 10:25:38 +05:30
Nick Rogers
f588e7783e Merge branch 'arc53:main' into main 2023-10-08 21:54:58 -07:00
Shuvadipta Das
f8ca6c019f Update My-AI-answers-questions-using-external-knowledge.md 2023-10-09 09:46:54 +05:30
Digvijay Shelar
f88806fc3c Update README.md 2023-10-09 07:26:34 +05:30
GH Action - Upstream Sync
ee0880fab7 Merge branch 'main' of https://github.com/arc53/DocsGPT 2023-10-09 00:30:35 +00:00
beKool.sh
261c674832 Update API-docs.md 2023-10-09 05:48:27 +05:45
Alex
e95bc82b8e Update Conversation.tsx 2023-10-08 23:30:21 +01:00
Alex
6d0cc49ecd Merge pull request #477 from staticGuru/input-hidden-issue-#474
[FIX] Question is hidden under the question input box
2023-10-08 23:27:45 +01:00
Alex
e108833db2 Merge branch 'main' into input-hidden-issue-#474 2023-10-08 23:19:35 +01:00
Alex
151fdb9bad Merge pull request #493 from HarshMN2345/patch-1
Update CONTRIBUTING.md
2023-10-08 22:57:19 +01:00
Alex
59ca8665fe Merge branch 'main' into patch-1 2023-10-08 22:56:20 +01:00
Alex
71c101b82e Merge pull request #519 from SoumyadiptoPal/newBranch2
Feature: Round Corners
2023-10-08 22:46:12 +01:00
Alex
860030824e Merge pull request #503 from ManishMadan2882/main
added the copy response feature
2023-10-08 22:40:51 +01:00
Alex
46c4bf6e94 Merge pull request #515 from timoransky/main
Fix: adjust left margin on content container
2023-10-08 22:32:55 +01:00
Alex
53ed6e54b5 Merge pull request #512 from drk1rd/main
Grammar and punctuations improved
2023-10-08 22:27:07 +01:00
ManishMadan2882
3197c356e9 UI corrections 2023-10-09 02:36:48 +05:30
Harsh Mahajan
cdad083d7f Update CONTRIBUTING.md 2023-10-09 01:21:36 +05:30
Alex
2e076ef3f4 Merge pull request #469 from 5h0ov/patch-1
Update How-to-train-on-other-documentation.md
2023-10-08 19:40:54 +01:00
Alex
46e3a27626 Merge pull request #509 from sanketmp/add-license
add license link to readme.md
2023-10-08 19:39:27 +01:00
Alex
1247867187 Merge pull request #480 from GuptaPratik02/improve-doc-readme-contributing-hacktoberfest-files
Improved docs - readme , contributing and hacktoberfest files
2023-10-08 19:36:55 +01:00
Soumyadipto Pal
b1f863cc4d Update Navigation.tsx 2023-10-08 23:59:08 +05:30
Soumyadipto Pal
823b41b7ec Merge branch 'main' into newBranch2 2023-10-08 23:53:35 +05:30
SoumyadiptoPal
16a2b3b19b Rounded the components to 3xl 2023-10-08 23:48:49 +05:30
Alex
0a2e899363 Merge pull request #506 from akshitarora921/fix/about-page
🪛 Fix: About us page margin
2023-10-08 18:50:08 +01:00
timoransky
65d431c7a0 fix: margin left on content container 2023-10-08 19:18:11 +02:00
staticGuru
6b617955b7 add the check mark logics 2023-10-08 22:31:17 +05:30
staticGuru
10cf0470cb add the conversation Tile 2023-10-08 22:09:38 +05:30
staticGuru
f91ca796de add the check Mark icons 2023-10-08 22:09:16 +05:30
Suryansh
7f1fb41d48 Update My-AI-answers-questions-using-external-knowledge.md 2023-10-08 22:08:55 +05:30
staticGuru
ceb9c70fba add the edit icons in the assets 2023-10-08 22:08:42 +05:30
staticGuru
5c9d11861e add the coversation tile in the chat section 2023-10-08 22:08:19 +05:30
Suryansh
706e6c01aa Update How-to-use-different-LLM.md 2023-10-08 22:08:18 +05:30
Suryansh
64cecb4931 Update How-to-train-on-other-documentation.md 2023-10-08 22:07:00 +05:30
Suryansh
31e0dfef76 Update Customising-prompts.md 2023-10-08 22:03:18 +05:30
Suryansh
dc85f93423 Update react-widget.md 2023-10-08 22:02:10 +05:30
Suryansh
4d5d407655 Update Chatwoot-extension.md 2023-10-08 22:00:46 +05:30
Suryansh
d2424ce540 Update API-docs.md 2023-10-08 21:59:27 +05:30
Suryansh
4d5de8176a Update Quickstart.md 2023-10-08 21:55:43 +05:30
Suryansh
c451d00eb4 Update Hosting-the-app.md 2023-10-08 21:52:52 +05:30
Harsh Mahajan
a8180bddad Update API-docs.md
done all the changes proposed in issue #508
2023-10-08 18:02:54 +05:30
Harsh Mahajan
e988364766 Update CONTRIBUTING.md
i made it easy to understand
2023-10-08 17:47:27 +05:30
staticGuru
396697ead2 Query overlay text input issues 2023-10-08 17:36:57 +05:30
Sanket Pol
2993bd8c05 add license link to readme.md 2023-10-08 17:02:44 +05:30
beKool.sh
fc50bb6e57 Add Backlink to Vector Stores 2023-10-08 17:10:33 +05:45
Pratik Gupta
a064066e42 Update README.md
Updated the DocsGPT link as mentioned by the maintainer.
2023-10-08 16:52:56 +05:30
Akshit Arora
a6783e537b fix css 2023-10-08 14:17:26 +05:30
Ayan Joshi
4b1dad96cd Commit 2023-10-08 12:42:56 +05:30
Digvijay Shelar
6758b51617 Update README.md 2023-10-08 11:20:28 +05:30
ManishMadan2882
54fdd2da57 UI checks 2023-10-08 06:26:08 +05:30
GH Action - Upstream Sync
3132a4965e Merge branch 'main' of https://github.com/arc53/DocsGPT 2023-10-08 00:32:29 +00:00
ManishMadan2882
7ee3f10a81 fix 2023-10-08 05:33:13 +05:30
ManishMadan2882
accd65a26a added the copy msg feature 2023-10-08 05:27:03 +05:30
Alex
e0ada7fc48 Merge pull request #496 from ankur0904/improve-docs
Make text bold
2023-10-08 00:33:28 +01:00
Alex
ad1401854c Merge pull request #448 from SoumyadiptoPal/newBranch
Updated navigation bar and also added icons
2023-10-08 00:27:58 +01:00
Alex
e18189caae Merge pull request #486 from krishvsoni/krishvsoni-patch-1
Krishvsoni patch 1
2023-10-08 00:19:25 +01:00
Alex
d601d35a21 Merge pull request #478 from adityagupta19/main
Fix "Sources" Feature doesn't look as intended
2023-10-08 00:10:13 +01:00
Alex
66fd402f00 Merge branch 'main' into main 2023-10-08 00:01:43 +01:00
Alex
835a04358c Merge pull request #492 from Ankit-Matth/change_color_of_buttons
Color of all buttons and elements changed from blue to #7D54D1
2023-10-07 23:56:55 +01:00
Alex
af9b4e448d Merge pull request #491 from daniel-shuy/bugfix/about-page-to-conversation-navigation
Fix About page to Conversation navigation
2023-10-07 23:52:40 +01:00
Alex
39e8ba42ff Merge pull request #454 from aindree-2005/patch-1
Update About.tsx
2023-10-07 23:29:25 +01:00
Alex
1e52c956a8 Merge pull request #462 from mohitd404/main
Update README.md -- Fixed #453
2023-10-07 23:18:20 +01:00
ankur0904
d261ed074e Make text bold 2023-10-07 23:54:47 +05:30
KRISH SONI
47f9be32ce git commit 2023-10-07 21:40:29 +05:30
Harsh Mahajan
a17390c157 Update CONTRIBUTING.md
i  made it easy to understand
2023-10-07 21:37:58 +05:30
Aditya Gupta
5ca5e0d00f Revert "synced the branch"
This reverts commit 0585fb4c80.
2023-10-07 21:22:48 +05:30
Ankit Matth
b0085f2741 Color of all buttons and elements changed from blue to #7D54D1 2023-10-07 20:42:32 +05:30
Daniel Shuy
b983095e13 Fix About page to Conversation navigation 2023-10-07 22:47:21 +08:00
KRISH SONI
dd6e018e46 Update README.md 2023-10-07 20:09:48 +05:30
KRISH SONI
6f8394a086 git commit 2023-10-07 20:00:17 +05:30
KRISH SONI
a0739a18e8 added one line for the venv documentation 2023-10-07 19:53:23 +05:30
KRISH SONI
27d33f015f added venv official Python documentation 2023-10-07 19:43:45 +05:30
Alex
ffb7ad1417 Merge pull request #463 from jbampton/remove-duplicate-words
Remove unneeded duplicate words
2023-10-07 14:12:00 +01:00
Aindree Chatterjee
97e6bab6e3 Update About.tsx 2023-10-07 18:41:39 +05:30
Aindree Chatterjee
b311b7620c Update About.tsx 2023-10-07 18:41:01 +05:30
Alex
25ec8fb2ab Merge pull request #465 from adarsh-jha-dev/patch-1
Add Feedback Feature to Conversation Module (Update conversationSlice.ts)
2023-10-07 13:58:37 +01:00
Alex
75100cd182 Merge pull request #473 from staticGuru/main
Hero section's figma UI changes
2023-10-07 13:33:38 +01:00
KRISH SONI
50d48ee3ec Added alternative virtual environment activation command for Windows users in README 2023-10-07 16:46:02 +05:30
KRISH SONI
0e330f983b Added alternative virtual environment activation command for Windows users in README 2023-10-07 16:44:55 +05:30
Mohit Dhote
9523a929af Merge branch 'main' into main 2023-10-07 15:23:53 +05:30
Mohit Dhote
3fcec069ed Update README.md 2023-10-07 15:14:04 +05:30
Mohit Dhote
7c2e72aebb Update README.md 2023-10-07 15:13:15 +05:30
Mohit Dhote
8d6fbddd67 Update README.md #453
made the required changes #453
2023-10-07 15:11:53 +05:30
Pratik Gupta
66b5ac8ff1 Update HACKTOBERFEST.md
Improvements
2023-10-07 14:51:22 +05:30
Pratik Gupta
e034fc12eb Update CONTRIBUTING.md
Fixed and added the necessary links in contributing file
2023-10-07 14:46:56 +05:30
Pratik Gupta
a8317ccacd Update README.md
Fixed and added the necessary links in README file
2023-10-07 14:37:54 +05:30
Alex
74376586a8 Update README.md 2023-10-07 09:51:41 +01:00
Pratik Gupta
ea49296cfe Update HACKTOBERFEST.md
Improved the HACKTOBERFEST.md file.
Fixed and added 7 links.
2023-10-07 14:15:17 +05:30
Aditya Gupta
992f817fef Fix: Sources feature doesn't look as intended 2023-10-07 13:24:12 +05:30
staticGuru
36528fceab add issue fixes 2023-10-07 12:20:47 +05:30
staticGuru
8323b8af4d Reset the input hidden issues 2023-10-07 12:13:05 +05:30
staticGuru
0e496181a1 Fix the question input hidden issue 2023-10-07 12:03:43 +05:30
staticGuru
f47fc7a484 add the additional div in the chat sections 2023-10-07 11:39:51 +05:30
Aditya Gupta
0585fb4c80 synced the branch 2023-10-07 11:21:55 +05:30
Nick Rogers
bdfcf6591e Fix missing documentation for using Llama_cpp 2023-10-06 21:20:40 -07:00
GH Action - Upstream Sync
cad54f0f07 Merge branch 'main' of https://github.com/arc53/DocsGPT 2023-10-07 01:15:01 +00:00
GH Action - Upstream Sync
a52ab1685e Merge branch 'main' of https://github.com/arc53/DocsGPT 2023-10-07 00:29:26 +00:00
Alex
3182816965 Merge pull request #472 from adityagupta19/main
Fix: Conversations tabs overlapping
2023-10-06 22:29:37 +01:00
Guruvignesh
a8da4b0162 Merge branch 'arc53:main' into main 2023-10-07 00:07:02 +05:30
staticGuru
ab7f6e8300 Change the hero section with figma style 2023-10-06 23:42:22 +05:30
Aditya Gupta
943bf477a0 Fix: Conversations tabs overlapping 2023-10-06 23:37:23 +05:30
Digvijay Shelar
168f4c0056 Merge branch 'main' into main 2023-10-06 22:41:28 +05:30
Digvijay Shelar
35fef11d2a Update README.md 2023-10-06 22:40:46 +05:30
Shuvadipta Das
425a8a6412 Update How-to-train-on-other-documentation.md 2023-10-06 21:39:18 +05:30
Alex
b64495f7a9 Merge pull request #467 from jbampton/add-pr-labeler
Add pull request labeler
2023-10-06 16:11:53 +01:00
Pavel
014861a7f2 Merge pull request #468 from arc53/bug/faiss-import
Fix faiss import bug
2023-10-06 18:09:31 +03:00
Alex
17edaa0e1f Update faiss.py 2023-10-06 16:05:10 +01:00
John Bampton
bbd0325c10 Add pull request labeler 2023-10-07 00:40:27 +10:00
Alex
316c276545 Update README.md 2023-10-06 15:29:17 +01:00
John Bampton
32ea0213f7 Remove unneeded duplicate words 2023-10-07 00:11:03 +10:00
Adarsh Jha
86c2f0716e Update conversationSlice.ts 2023-10-06 19:36:48 +05:30
Pavel
68b8d7d7f2 Merge pull request #464 from arc53/bug/fix-tests
Fix tests on sagemaker.py
2023-10-06 17:02:54 +03:00
Alex
43a22f84d9 Update sagemaker.py 2023-10-06 14:43:05 +01:00
Alex
b3a0368b95 Update app.py 2023-10-06 13:54:03 +01:00
Alex
cd79330c4c Merge pull request #449 from arc53/feature/sage-streaming
sagemaker streaming
2023-10-06 13:49:32 +01:00
mohitd404
245e09c723 Update README.md -- Fixed ##453
 Added code of conduct section to documentation.
 Added license section into the repository.
  Fixed some Typos.
2023-10-06 18:01:50 +05:30
Alex
495728593f sagemaker fixes + test 2023-10-06 13:22:51 +01:00
Aditya Aryaman Das
e9c4b0dc01 Updated README.md 2023-10-06 17:24:49 +05:30
Alex
9942bf2124 Merge pull request #458 from staticGuru/main
[Fix] Source selection overflow issue
2023-10-06 12:37:14 +01:00
staticGuru
0a8ba068c4 remove the unwanted whitespace 2023-10-06 17:04:25 +05:30
staticGuru
a2bb70aaec Source selection overflow issues 2023-10-06 16:51:35 +05:30
Alex
5ed25d8bcb Merge pull request #447 from byt3h3ad/byt3h3ad-patch-1
improved CONTRIBUTING.MD, README.MD, HACKTOBERFEST.MD
2023-10-06 12:03:27 +01:00
Alex
cafc068c39 Merge branch 'main' into byt3h3ad-patch-1 2023-10-06 12:03:20 +01:00
Alex
b8dde0767b Update docker-compose.yaml 2023-10-06 11:57:08 +01:00
Alex
f8e5e3b3c0 Merge pull request #445 from amoghak-ds/main
Update CODE_OF_CONDUCT.md
2023-10-06 11:53:03 +01:00
Aindree Chatterjee
edc19e99a9 Merge branch 'arc53:main' into patch-1 2023-10-06 16:20:11 +05:30
Alex
2b0b3827ab Merge pull request #443 from amoghak-ds/patch-1
Update app.py
2023-10-06 11:49:16 +01:00
Aindree Chatterjee
8afe5a0087 Update About.tsx 2023-10-06 16:19:05 +05:30
Alex
0ecc53f3b6 Merge pull request #456 from LunarMarathon/fix-455-bitsandbytes
correct wrapper, typos; add links
2023-10-06 11:21:11 +01:00
LunarMarathon
b3f2827961 correct wrapper, typos; add links
Signed-off-by: LunarMarathon <lmaytan24@gmail.com>
2023-10-06 15:11:22 +05:30
Aindree Chatterjee
9c96a4d81b Update About.tsx
Text decorations added
2023-10-06 12:28:32 +05:30
Alex
4f5e363452 sagemaker streaming 2023-10-06 01:52:29 +01:00
Alex
92572ff919 Merge pull request #442 from ratishjain12/fixed-homepage
fixed homepage
2023-10-06 01:32:34 +01:00
Ratish jain
39ddaf49be Update Hero.tsx 2023-10-06 05:37:44 +05:30
Alex
627dc2d4a0 Merge pull request #441 from ka1bi4/update/documentation-update-formatting
Improved docs readability and fix some typo.
2023-10-05 23:33:00 +01:00
Soumyadipto Pal
42739bbb61 Merge branch 'main' into newBranch 2023-10-06 02:36:49 +05:30
SoumyadiptoPal
261c9eefe1 Updated navigation bar and also added discord and github icons 2023-10-06 02:10:21 +05:30
Adhiraj
e21e4d2b16 Update README.md 2023-10-06 02:05:11 +05:30
Adhiraj
8b6b8f0c53 Update CONTRIBUTING.md 2023-10-06 01:40:03 +05:30
Adhiraj
5a9feb4411 Update HACKTOBERFEST.md 2023-10-06 01:31:37 +05:30
Amogha Kancharla
f16128da09 Update CODE_OF_CONDUCT.md
I've improved the CODE_OF_CONDUCT.md file and fixed some typos and errors. Kindly review it.
2023-10-06 00:07:13 +05:30
Amogha Kancharla
48f9997ea9 Update app.py
This version of code uses a concise syntax and reduces redundancy.
2023-10-05 23:35:47 +05:30
ratishjain
f0e87094d6 fixed homepage 2023-10-05 23:24:09 +05:30
ratishjain
e0882e9e04 fixed homepage 2023-10-05 23:16:54 +05:30
Roman Zhukov
d37885ea88 Update doc formatting and fix some spelling. 2023-10-05 20:27:48 +03:00
Alex
d13e5e7e3f Merge pull request #435 from Ankit-Matth/changing_discord_github_icons
I have changed icons & text for Discord and Github in the left sidebar.
2023-10-05 16:35:53 +01:00
Alex
aa9a024ee1 Merge branch 'main' into changing_discord_github_icons 2023-10-05 16:33:18 +01:00
Digvijay Shelar
5bbf6d2ae9 Update README.md 2023-10-05 20:38:02 +05:30
Alex
30299a9f04 Merge pull request #434 from QuantuM410/quantum410/frontend-placeholder
Added css attribute placeholder for search div
2023-10-05 15:55:00 +01:00
Alex
23d7fe936d Merge pull request #432 from arc53/dependabot/pip/application/langchain-0.0.308
Bump langchain from 0.0.263 to 0.0.308 in /application
2023-10-05 15:46:23 +01:00
Alex
b50c052222 Merge pull request #399 from arc53/dependabot/npm_and_yarn/extensions/web-widget/postcss-8.4.31
Bump postcss from 8.4.23 to 8.4.31 in /extensions/web-widget
2023-10-05 15:41:28 +01:00
dependabot[bot]
ef9e9809e2 Bump langchain from 0.0.263 to 0.0.308 in /application
Bumps [langchain](https://github.com/langchain-ai/langchain) from 0.0.263 to 0.0.308.
- [Release notes](https://github.com/langchain-ai/langchain/releases)
- [Commits](https://github.com/langchain-ai/langchain/compare/v0.0.263...v0.0.308)

---
updated-dependencies:
- dependency-name: langchain
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-10-05 14:38:51 +00:00
Alex
f139c3268b Merge pull request #431 from arc53/dependabot/pip/scripts/langchain-0.0.308
Bump langchain from 0.0.252 to 0.0.308 in /scripts
2023-10-05 15:37:53 +01:00
dependabot[bot]
e869bfd991 Bump postcss from 8.4.23 to 8.4.31 in /extensions/web-widget
Dependabot couldn't find the original pull request head commit, b4ce5c5d7b37c529245006df293f65d8065dd513.
2023-10-05 14:36:56 +00:00
Alex
d5309fcaf5 Merge pull request #398 from arc53/dependabot/npm_and_yarn/extensions/chrome/postcss-8.4.31
Bump postcss from 8.4.21 to 8.4.31 in /extensions/chrome
2023-10-05 15:36:41 +01:00
Alex
c4fc49553c Merge pull request #396 from arc53/dependabot/pip/application/pillow-10.0.1
Bump pillow from 9.4.0 to 10.0.1 in /application
2023-10-05 15:36:07 +01:00
Alex
75704899a7 Merge pull request #395 from arc53/dependabot/npm_and_yarn/extensions/react-widget/postcss-8.4.31
Bump postcss from 8.4.29 to 8.4.31 in /extensions/react-widget
2023-10-05 15:35:48 +01:00
QuantuM410
70aa3b1ff1 Added css attribute placeholder for search div 2023-10-05 20:01:30 +05:30
Ankit Matth
6154a8169b Replace Discord & Github Icons 2023-10-05 19:53:37 +05:30
dependabot[bot]
cf0173e079 Bump langchain from 0.0.252 to 0.0.308 in /scripts
Bumps [langchain](https://github.com/langchain-ai/langchain) from 0.0.252 to 0.0.308.
- [Release notes](https://github.com/langchain-ai/langchain/releases)
- [Commits](https://github.com/langchain-ai/langchain/compare/v0.0.252...v0.0.308)

---
updated-dependencies:
- dependency-name: langchain
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-10-05 13:47:34 +00:00
Alex
cc62fc6222 Merge pull request #429 from shelar1423/main
fix: Typo in Readme.md
2023-10-05 14:00:33 +01:00
Digvijay Shelar
6cdadf1b37 Update README.md 2023-10-05 17:19:21 +05:30
Alex
dc90a66a96 Merge pull request #427 from ArnabBCA/main
Fixed Overflowing Text messages with long strings without spaces
2023-10-05 11:10:39 +01:00
Alex
4b629d20cf Merge pull request #425 from Archit-Kohli/patch-2
Update huggingface.py
2023-10-05 11:03:27 +01:00
Arnab Ghosh
437bd13fd0 Fixed Overflowing Text message when user passed a very long string without spaces 2023-10-05 14:44:18 +05:30
Archit-Kohli
7ce1dc9069 Update huggingface.py add import torch statement
Added import torch statement
2023-10-05 14:27:15 +05:30
Archit-Kohli
5b4e517d9d Update huggingface.py
Added quantization support using bitsandbytes
2023-10-05 11:03:51 +05:30
Alex
6c3ed5e533 Merge pull request #417 from eltociear/patch-2
Update How-to-use-different-LLM.md
2023-10-05 00:14:43 +01:00
Alex
ec2762c31a Merge pull request #384 from ArnabBCA/main
Fixed Empty Spaces Passed in the Input Field
2023-10-05 00:10:51 +01:00
Alex
29f3158b61 Merge pull request #407 from Akash190104/patch-2
[Docs] Update Quickstart.md
2023-10-04 23:58:41 +01:00
Alex
d83b7276fd Merge pull request #406 from Akash190104/patch-1
[DOCS] Update Hosting-the-app.md
2023-10-04 23:57:19 +01:00
Alex
1336010bb2 Update .env_sample 2023-10-04 23:54:53 +01:00
Alex
5cb3df6db1 Merge pull request #420 from jerempy/feature/useMediaQuery
useMediaQuery
2023-10-04 18:50:42 +01:00
jerempy
4be0c1c0eb delete unused test 2023-10-04 13:41:43 -04:00
jerempy
33e5e74228 custom hook 2023-10-04 13:12:52 -04:00
Ikko Eltociear Ashimine
2b06989372 Update How-to-use-different-LLM.md
Huggingface -> Hugging Face
2023-10-05 01:24:46 +09:00
Alex
cd4da2aca3 Merge pull request #402 from jbampton/fix-spelling
react-widget: fix spelling; `atrribute`
2023-10-04 16:38:28 +01:00
Alex
b335951862 Merge pull request #408 from Shrit1401/main
Added Contributors Images in Readme.md
2023-10-04 16:35:23 +01:00
Arnab Ghosh
e95e084956 Fixed Empty Spaces Passed in the Input Field 2023-10-04 20:54:25 +05:30
Alex
b7d569de98 Update README.md 2023-10-04 14:01:23 +01:00
Alex
8320cca5cd minor fixes on deleting 2023-10-04 13:57:40 +01:00
Shrit Shrivastava
bad5fec0f1 Added Contributors Images in Readme.md 2023-10-04 18:16:53 +05:30
Akash Kundu
200a3b65ee [Docs] Update Quickstart.md 2023-10-04 18:06:44 +05:30
Akash Kundu
d9fc2a93cc Update Hosting-the-app.md
There were a lot of grammatical errors and typos that I fixed to improve the readability of the document.
2023-10-04 17:56:25 +05:30
Alex
94f3533c29 Update README.md 2023-10-04 12:33:58 +01:00
John Bampton
77f7ad309e react-widget: fix spelling; atrribute 2023-10-04 10:12:46 +10:00
dependabot[bot]
cefd270837 Bump postcss from 8.4.21 to 8.4.31 in /extensions/chrome
Bumps [postcss](https://github.com/postcss/postcss) from 8.4.21 to 8.4.31.
- [Release notes](https://github.com/postcss/postcss/releases)
- [Changelog](https://github.com/postcss/postcss/blob/main/CHANGELOG.md)
- [Commits](https://github.com/postcss/postcss/compare/8.4.21...8.4.31)

---
updated-dependencies:
- dependency-name: postcss
  dependency-type: indirect
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-10-03 22:52:59 +00:00
dependabot[bot]
037e68a376 Bump postcss from 8.4.29 to 8.4.31 in /extensions/react-widget
Bumps [postcss](https://github.com/postcss/postcss) from 8.4.29 to 8.4.31.
- [Release notes](https://github.com/postcss/postcss/releases)
- [Changelog](https://github.com/postcss/postcss/blob/main/CHANGELOG.md)
- [Commits](https://github.com/postcss/postcss/compare/8.4.29...8.4.31)

---
updated-dependencies:
- dependency-name: postcss
  dependency-type: direct:development
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-10-03 22:52:56 +00:00
Alex
4da0785494 Merge pull request #397 from arc53/dependabot/npm_and_yarn/frontend/postcss-8.4.31
Bump postcss from 8.4.21 to 8.4.31 in /frontend
2023-10-03 23:52:19 +01:00
dependabot[bot]
53171bafec Bump pillow from 9.4.0 to 10.0.1 in /application
Bumps [pillow](https://github.com/python-pillow/Pillow) from 9.4.0 to 10.0.1.
- [Release notes](https://github.com/python-pillow/Pillow/releases)
- [Changelog](https://github.com/python-pillow/Pillow/blob/main/CHANGES.rst)
- [Commits](https://github.com/python-pillow/Pillow/compare/9.4.0...10.0.1)

---
updated-dependencies:
- dependency-name: pillow
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-10-03 22:46:06 +00:00
dependabot[bot]
16fe77e472 Bump postcss from 8.4.21 to 8.4.31 in /frontend
Bumps [postcss](https://github.com/postcss/postcss) from 8.4.21 to 8.4.31.
- [Release notes](https://github.com/postcss/postcss/releases)
- [Changelog](https://github.com/postcss/postcss/blob/main/CHANGELOG.md)
- [Commits](https://github.com/postcss/postcss/compare/8.4.21...8.4.31)

---
updated-dependencies:
- dependency-name: postcss
  dependency-type: direct:development
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-10-03 22:46:06 +00:00
Alex
9f147e5b6d Merge pull request #393 from arc53/dependabot/pip/scripts/pillow-10.0.1
Bump pillow from 10.0.0 to 10.0.1 in /scripts
2023-10-03 23:45:37 +01:00
Alex
c43106a744 Merge pull request #392 from jbampton/remove-unneeded-comments
tests(python): remove unneeded comments
2023-10-03 23:43:24 +01:00
Alex
a73eb0377d Merge pull request #389 from Cioraz/fixed_sumbit_to_submit
Corrected Sumbit to Submit
2023-10-03 23:36:00 +01:00
Alex
e05514b455 Merge pull request #372 from siiddhantt/main
Updated sources section according to Figma design
2023-10-03 23:21:32 +01:00
dependabot[bot]
901c7be9a8 Bump pillow from 10.0.0 to 10.0.1 in /scripts
Bumps [pillow](https://github.com/python-pillow/Pillow) from 10.0.0 to 10.0.1.
- [Release notes](https://github.com/python-pillow/Pillow/releases)
- [Changelog](https://github.com/python-pillow/Pillow/blob/main/CHANGES.rst)
- [Commits](https://github.com/python-pillow/Pillow/compare/10.0.0...10.0.1)

---
updated-dependencies:
- dependency-name: pillow
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-10-03 20:40:53 +00:00
John Bampton
e503cc3003 tests(python): remove unneeded comments 2023-10-04 06:06:20 +10:00
Alex
2ff777acb7 Merge pull request #390 from jbampton/fix-spelling
misc: fix spelling
2023-10-03 20:25:46 +01:00
John Bampton
034d73a4eb misc: fix spelling 2023-10-04 05:18:15 +10:00
Siddhant Rai
aec55e50f9 Merge branch 'arc53:main' into main 2023-10-03 23:32:43 +05:30
siiddhantt
1f356a67b2 fix: added conditions for error state 2023-10-03 23:32:08 +05:30
Cioraz
0f60cd480d Correct Sumbit to Submit 2023-10-03 22:27:58 +05:30
Cioraz
316b8c66db Correct Sumbit to Submit 2023-10-03 22:26:53 +05:30
Alex
744d4ebbaf Update bug_report.yml 2023-10-03 15:14:44 +01:00
Alex
005deaccc8 Merge pull request #381 from adeyinkaezra123/issuetemplates
feat: create issue template forms to ensure consistency in issue and pr formats
2023-10-03 15:04:05 +01:00
Ezra Adeyinka
b6f78ce1af fix(bug_report_template): add environments and variable constructs 2023-10-03 14:39:56 +01:00
Ezra Adeyinka
932b504d82 fix(bug_report_template): append working code of conduct url 2023-10-03 14:22:53 +01:00
Ezra Adeyinka
1cca46cf7b ci(root): add pull request markdown template 2023-10-03 13:04:32 +01:00
Ezra Adeyinka
a8f6d2adf0 ci(root): add feature request issue form template directive 2023-10-03 13:02:24 +01:00
Ezra Adeyinka
203de18053 ci(root): add bug report issue form template directive 2023-10-03 12:47:10 +01:00
Alex
ee12b4164b Merge pull request #337 from arc53/dependabot/pip/scripts/cryptography-41.0.4
Bump cryptography from 41.0.3 to 41.0.4 in /scripts
2023-10-03 11:12:21 +01:00
Alex
b38459439d Merge pull request #338 from arc53/dependabot/pip/application/cryptography-41.0.4
Bump cryptography from 41.0.3 to 41.0.4 in /application
2023-10-03 11:12:02 +01:00
Alex
a2eddb3580 Merge pull request #373 from arc53/dependabot/pip/application/urllib3-1.26.17
Bump urllib3 from 1.26.14 to 1.26.17 in /application
2023-10-03 11:07:34 +01:00
Alex
18adbc6bf0 Merge pull request #377 from prithvi2k2/main
Removed redundant files as discussed at #376
2023-10-03 11:06:52 +01:00
Alex
c1ccef25a3 Merge pull request #375 from chriscarnold/github-star-badge-readme
Fixed GihHub badges in README.md links
2023-10-03 11:02:04 +01:00
Alex
b1bea73efb Merge pull request #374 from Smartmind12/patch-2
Adding tab to documentation in Navigation bar
2023-10-03 10:57:56 +01:00
Alex
4175d29056 hotfix 2023-10-03 10:50:49 +01:00
Alex
46c78c33cf Merge pull request #378 from arc53/feature/holopin
Feature/holopin
2023-10-03 09:49:38 +01:00
Alex
3fe5a41433 Update HACKTOBERFEST.md 2023-10-03 09:49:03 +01:00
Alex
f2b1f95521 Update holopin.yml 2023-10-03 09:46:25 +01:00
prithvi2k2
94f81caf28 Removed redundant files as discussed at #376 2023-10-03 14:11:52 +05:30
chriscarnold
669a4a299c Fixed GihHub badges in README.md links 2023-10-03 09:05:14 +01:00
siiddhantt
afff55045f fix: css changes 2023-10-03 10:47:15 +05:30
Siddhant Rai
4fbcd2ba5d Merge branch 'arc53:main' into main 2023-10-03 10:45:41 +05:30
Utsav Paul
016295dfee Adding tab to documentation in Navigation bar 2023-10-03 10:33:02 +05:30
dependabot[bot]
6d2bc2929a Bump urllib3 from 1.26.14 to 1.26.17 in /application
Bumps [urllib3](https://github.com/urllib3/urllib3) from 1.26.14 to 1.26.17.
- [Release notes](https://github.com/urllib3/urllib3/releases)
- [Changelog](https://github.com/urllib3/urllib3/blob/main/CHANGES.rst)
- [Commits](https://github.com/urllib3/urllib3/compare/1.26.14...1.26.17)

---
updated-dependencies:
- dependency-name: urllib3
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-10-03 02:47:59 +00:00
Alex
23a5e566f2 Update Quickstart.md 2023-10-03 00:39:28 +01:00
Alex
4b387961a4 Update Quickstart.md 2023-10-03 00:39:12 +01:00
siiddhantt
1765a8a7f9 fix: fixes according to figma design 2023-10-02 23:22:04 +05:30
siiddhantt
837a5b52a7 Merge branch 'main' of github.com:siiddhantt/DocsGPT 2023-10-02 22:42:20 +05:30
siiddhantt
180c4e855e feat: updated sources section design 2023-10-02 22:37:12 +05:30
Alex
01d2af9961 Update favicon.ico
favicon new
2023-10-02 15:21:36 +01:00
Alex
7f3cc6269b holopin init 2023-10-02 15:14:19 +01:00
Alex
dbc0d54491 Update index.mdx 2023-10-02 10:04:08 +01:00
Alex
f843f5ae9d Update index.mdx 2023-10-02 10:03:34 +01:00
Alex
962cb290e4 Merge pull request #366 from Smartmind12/patch-1
Update index.mdx
2023-10-02 09:53:39 +01:00
Alex
7138655dd1 Update HACKTOBERFEST.md 2023-10-02 09:47:39 +01:00
Utsav Paul
93acfc2e38 Update index.mdx
Updating the Documentation Landing page to make it more userfriendly.
2023-10-02 11:16:44 +05:30
Alex
91878c4591 Update README.md 2023-10-02 00:31:24 +01:00
Alex
05ec1216e0 Merge pull request #363 from hirenchauhan2/feat/ui/170
feat(ui): add scroll to bottom button
2023-10-01 21:59:22 +01:00
Alex
af0e6481f8 Update Conversation.tsx 2023-10-01 21:57:12 +01:00
Alex
11a745c4d9 Merge pull request #364 from arc53/bug/fix-test
Fix broken test
2023-10-01 21:45:31 +01:00
Alex
2393da4425 Update README.md 2023-10-01 21:44:54 +01:00
Alex
95ab08e02d Update test_celery.py 2023-10-01 21:34:57 +01:00
Alex
95fdedf12e Merge pull request #361 from Ayush-Prabhu/patch-1
Grammatical corrections
2023-10-01 20:58:09 +01:00
Alex
c73dd776db Merge pull request #362 from arc53/feature/startup-script-cpu-inference
script + cpu optimisations
2023-10-01 20:11:42 +01:00
Alex
891e5fea3f Update README.md 2023-10-01 20:10:41 +01:00
Alex
bb2f6f23b5 Update README.md 2023-10-01 20:09:15 +01:00
Alex
cd9b03bdb9 celery syncs 2023-10-01 20:05:13 +01:00
Alex
a619269502 celery bugs 2023-10-01 19:55:11 +01:00
Hiren Chauhan
153b5c028b feat(ui): add scroll to bottom button
This will show the scroll to bottom button when user scrolls to top from the last message.

Closes #170
2023-10-02 00:15:31 +05:30
Alex
9a33bf2210 script + cpu optimisations 2023-10-01 19:16:13 +01:00
Ayush-Prabhu
34b4cd2231 Grammatical corrections
Corrected grammatical errors and spelling errors in : docs/pages/Guides/My-AI-answers-questions-using-external-knowledge.md
2023-10-01 23:36:50 +05:30
Alex
6045cbbc62 Merge pull request #355 from arc53/feature/cpu-llm
llama-cpp local
2023-10-01 17:55:26 +01:00
Alex
9bbf4044e0 script 2023-10-01 17:20:47 +01:00
Alex
fcf8a64d91 Merge pull request #360 from jbampton/fix-spelling
Fix spelling
2023-10-01 17:09:53 +01:00
John Bampton
2c6ab18e41 Fix spelling 2023-10-02 01:25:23 +10:00
Alex
2fea294b13 Update settings.py 2023-10-01 11:28:06 +01:00
Pavel
b47ecab1a9 llama-cpp local 2023-09-30 23:38:48 +04:00
Pavel
b86c294250 Merge pull request #354 from arc53/featue/elasticsearch
working es
2023-09-30 17:37:37 +03:00
Alex
3eacfb91aa fix314 2023-09-30 15:32:37 +01:00
Alex
94164c2a71 Merge branch 'main' into featue/elasticsearch 2023-09-30 15:30:23 +01:00
Alex
d85eb83ea2 elastic search fixes 2023-09-30 15:25:31 +01:00
Alex
b2002639db Merge pull request #353 from arc53/bug/fix-aes-pdf
Update requirements.txt
2023-09-29 17:34:17 +01:00
Alex
347cfe253f elastic2 2023-09-29 17:17:48 +01:00
Pavel
833e1836e1 Merge pull request #352 from arc53/feature/aws-sagemaker-inference
sagemaker + llm creator class
2023-09-29 17:42:54 +03:00
Alex
e4be38b9f7 sagemaker + llm creator class 2023-09-29 01:09:01 +01:00
Alex
783e7f6939 working es 2023-09-29 00:32:19 +01:00
Alex
c1c54f4848 Update README.md 2023-09-28 16:07:50 +01:00
Alex
86be6be2d2 Update Dockerfile 2023-09-28 15:30:47 +01:00
Alex
35a63e867a Merge pull request #345 from beardcodes/patch-1
Update Dockerfile
2023-09-28 15:30:19 +01:00
Alex
9c12a417ee Update README.md 2023-09-28 15:22:56 +01:00
Alex
32a019c0d6 Update requirements.txt 2023-09-27 22:39:48 +01:00
Pavel
b7e4a3c99e Merge pull request #348 from arc53/feature/better-structure
Feature/better structure
2023-09-27 20:18:09 +03:00
Alex
039062d071 ruff fix 2023-09-27 18:10:26 +01:00
Alex
83ae3e8371 more ruff fixes 2023-09-27 18:04:07 +01:00
Alex
852de8bdfc ruff linting 2023-09-27 18:01:40 +01:00
Alex
b8acb860aa some tests 2023-09-27 17:54:57 +01:00
Alex
e6849b85d1 Create huggingface.py 2023-09-27 17:02:47 +01:00
Alex
8fa9657ba6 working full 2023-09-27 16:25:57 +01:00
Zakarya El Quaroui
04b038960b Update Dockerfile
The current node version is vulnerable to buffer overflow. 

CVE-2022-3602
PUBLISHED
View JSON
X.509 Email Address 4-byte Buffer Overflow
Important CVE JSON 5 Information
Assigner: Openssl

Published: 2022-11-01Updated: 2022-11-03
A buffer overrun can be triggered in X.509 certificate verification, specifically in name constraint checking. Note that this occurs after certificate chain signature verification and requires either a CA to have signed the malicious certificate or for the application to continue certificate verification despite failure to construct a path to a trusted issuer. An attacker can craft a malicious email address to overflow four attacker-controlled bytes on the stack. This buffer overflow could result in a crash (causing a denial of service) or potentially remote code execution. Many platforms implement stack overflow protections which would mitigate against the risk of remote code execution. The risk may be further mitigated based on stack layout for any given platform/compiler. Pre-announcements of CVE-2022-3602 described this issue as CRITICAL. Further analysis based on some of the mitigating factors described above have led this to be downgraded to HIGH. Users are still encouraged to upgrade to a new version as soon as possible. In a TLS client, this can be triggered by connecting to a malicious server. In a TLS server, this can be triggered if the server requests client authentication and a malicious client connects. Fixed in OpenSSL 3.0.7 (Affected 3.0.0,3.0.1,3.0.2,3.0.3,3.0.4,3.0.5,3.0.6).
2023-09-27 17:08:44 +08:00
Alex
52507a5a95 Merge pull request #342 from arc53/hacktoberfest 2023-09-26 18:36:14 +01:00
Alex
d8505ba2ab Update README.md 2023-09-26 15:14:26 +01:00
Alex
fa26c0997e Update index.mdx 2023-09-26 15:07:42 +01:00
Alex
5a0aadd2ae Hacktoberfest info 2023-09-26 13:48:57 +01:00
Alex
025549ebf8 fixes to make it work 2023-09-26 13:00:17 +01:00
Alex
e85a583f0a testings 2023-09-26 10:03:22 +01:00
Alex
f7244ddb7a Merge pull request #340 from DenyTwice/main
UI Improvements, implements task 3 in issue #279
2023-09-24 11:13:27 +01:00
DenyTwice
d983a519e3 Uncomments selectDocsModal, removes redundant styles 2023-09-23 21:43:16 +05:30
DenyTwice
ae01070b8f Design consistency changes, fixes arrow icon positioning in source docs dropdown 2023-09-23 21:31:05 +05:30
dependabot[bot]
0ffb40f4c1 Bump cryptography from 41.0.3 to 41.0.4 in /application
Bumps [cryptography](https://github.com/pyca/cryptography) from 41.0.3 to 41.0.4.
- [Changelog](https://github.com/pyca/cryptography/blob/main/CHANGELOG.rst)
- [Commits](https://github.com/pyca/cryptography/compare/41.0.3...41.0.4)

---
updated-dependencies:
- dependency-name: cryptography
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-09-21 21:13:00 +00:00
dependabot[bot]
8bcffb4ad5 Bump cryptography from 41.0.3 to 41.0.4 in /scripts
Bumps [cryptography](https://github.com/pyca/cryptography) from 41.0.3 to 41.0.4.
- [Changelog](https://github.com/pyca/cryptography/blob/main/CHANGELOG.rst)
- [Commits](https://github.com/pyca/cryptography/compare/41.0.3...41.0.4)

---
updated-dependencies:
- dependency-name: cryptography
  dependency-type: direct:production
...

Signed-off-by: dependabot[bot] <support@github.com>
2023-09-21 21:06:30 +00:00
Alex
b2118602d9 Merge pull request #335 from B2o5T/patch-1
fix syntax highlightning
2023-09-16 09:49:03 +01:00
Dimitri POSTOLOV
9303f3b47b Update API-docs.md 2023-09-16 02:18:01 +02:00
Alex
e5c43cfc4b Merge pull request #334 from arc53/support-for-docx
Include docx files in the frontend
2023-09-15 11:28:56 +01:00
Alex
45fc08e221 Update Upload.tsx 2023-09-15 11:28:23 +01:00
Alex
67e8511106 Update Upload.tsx 2023-09-15 11:27:08 +01:00
Pavel
4f7fd0a62b Merge pull request #333 from arc53/feature/update-guides
updated deployment and created react widget guide
2023-09-15 13:11:30 +03:00
Alex
88fe454962 removed unecessary comma 2023-09-15 11:08:21 +01:00
Alex
26f7a9be0a updated deployment and create react widget guide 2023-09-15 11:00:59 +01:00
Alex
9256926bb7 Update README.md 2023-09-14 22:22:28 +01:00
Alex
2a83318739 updates modal 2023-09-13 14:11:32 +01:00
Pavel
d6e2535a5e Merge pull request #330 from arc53/feature/better-widget
Feature/better widget
2023-09-12 20:05:01 +03:00
Alex
2bffb7e22c update widgets 2023-09-12 17:44:40 +01:00
Alex
24a162cf86 use all states 2023-09-12 17:43:41 +01:00
Alex
f3104f3bc4 different source docs 2023-09-12 17:37:26 +01:00
Alex
45f1bf6709 widget final 2023-09-12 17:36:41 +01:00
Alex
40b2590815 different imports 2023-09-12 17:25:08 +01:00
Alex
dd9ab46b5c Update theme.config.jsx 2023-09-12 17:21:32 +01:00
Alex
c2aeadae33 Update theme.config.jsx 2023-09-12 17:19:18 +01:00
Alex
1bd9759ab7 update package 2023-09-12 17:13:34 +01:00
Alex
dcdbb05168 Update theme.config.jsx 2023-09-12 17:00:45 +01:00
Alex
ae117c47e9 widget everywhere 2023-09-12 16:43:47 +01:00
Alex
7f7856f0e4 Local storage sync 2023-09-12 16:39:09 +01:00
Alex
aa7b7c8619 Update docs 2023-09-12 15:48:52 +01:00
Alex
ee0cbff245 cleanup 2023-09-12 15:42:31 +01:00
Alex
c2c18b25d2 widget 0.2.0 2023-09-12 15:41:05 +01:00
Alex
816c7c95ed react-widget 2023-09-12 14:01:12 +01:00
Alex
cb5d65d11a widget init 2023-09-08 13:30:08 +01:00
Alex
75f3f43ba0 Merge pull request #327 from larinam/patch-2 2023-09-08 01:31:55 +01:00
Alex
9a521355ed Merge pull request #326 from larinam/remove-static 2023-09-08 01:30:59 +01:00
Anton Larin
47bfdf0710 Extended info on .env 2023-09-07 21:16:03 +02:00
Anton Larin
e1b49c3fb4 remove old static resources from the Flask application, forgotten leftover. 2023-09-07 18:32:45 +02:00
Alex
374dffc5fa little fix 2023-09-07 12:43:59 +01:00
Alex
4f735a5d11 Nextra docs 2023-09-07 12:36:39 +01:00
Alex
94738d8fc4 Merge pull request #325 from larinam/remove-static
remove old static resources from the Flask application, update the ro…
2023-09-07 09:51:33 +01:00
Anton Larin
adb4bfa10b remove old static resources from the Flask application, update the routing in app.py 2023-09-07 10:19:58 +02:00
Alex
48e6bbdc97 Merge pull request #322 from larinam/patch-1
Update CONTRIBUTING.md - information about running unit tests
2023-09-05 10:01:37 +01:00
Anton Larin
b54d6fea44 Update CONTRIBUTING.md - information about running unit tests 2023-09-05 06:31:27 +02:00
Alex
4462e6339d Merge pull request #320 from larinam/test-codecov
add simple test to make a PR to check CodeCov
2023-09-04 18:38:46 +01:00
Anton Larin
c1581b69f4 small optimization 2023-09-04 19:32:56 +02:00
Alex
14284e0cc7 Update test_app.py 2023-09-04 18:25:41 +01:00
Anton Larin
de40e733ec add simple test to make a PR to check CodeCov 2023-09-04 19:13:51 +02:00
Alex
9d91b6f780 Merge pull request #315 from arc53/codecov-integration
Create codecov.yml
2023-09-04 16:24:02 +01:00
Alex
6a8b49f9c4 Create codecov.yml 2023-09-04 14:48:20 +01:00
Alex
445a8a5647 Merge pull request #313 from arc53/codecov-integration
Update pytest.yml
2023-09-04 14:46:59 +01:00
Alex
83ce4a538a Update pytest.yml 2023-09-04 14:23:44 +01:00
Alex
35a19d2007 Update .env-template 2023-08-31 13:48:33 +01:00
Alex
505e12c5ea Merge pull request #306 from larinam/pytest-introduction
Adapt documentation to existing tests.
2023-08-23 09:37:57 +01:00
Alex
b2bfd7f23a Update README.md 2023-08-22 09:32:20 +01:00
Alex
cdb96e715d Update README.md 2023-08-21 22:19:47 +01:00
Alex
b3e5f09e3b Merge pull request #308 from larinam/revert-pytest-introduction
revert introduction of the coverage note addition to pull requests as…
2023-08-21 22:18:47 +01:00
Alex
db542d668a Update README.md 2023-08-21 22:16:50 +01:00
Anton Larin
a8a79a55a4 revert introduction of the coverage note addition to pull requests as it doesn't work for pull requests from public forks. see GitHub documentation: https://docs.github.com/en/actions/security-guides/automatic-token-authentication#permissions-for-the-github_token 2023-08-21 21:24:29 +02:00
Anton Larin
47f62a87a7 Revert "experiment with permissions"
This reverts commit 44f353861a.
2023-08-21 20:19:37 +02:00
Anton Larin
44f353861a experiment with permissions 2023-08-21 13:48:18 +02:00
Anton Larin
a2ef84a4a0 Adapt documentation to existing tests. 2023-08-18 17:43:17 +02:00
Alex
12ac20ec43 Merge pull request #304 from larinam/pytest-introduction
count test coverage
2023-08-17 16:17:57 +01:00
Anton Larin
ecfbc7b9fd count coverage 2023-08-16 16:35:48 +02:00
Alex
ba2fe0fb1f Update README.md 2023-08-15 14:54:19 +01:00
239 changed files with 21530 additions and 4302 deletions

View File

@@ -1,7 +1,8 @@
OPENAI_API_KEY=<LLM api key (for example, open ai key)>
EMBEDDINGS_KEY=<LLM embeddings api key (for example, open ai key)>
API_KEY=<LLM api key (for example, open ai key)>
LLM_NAME=docsgpt
VITE_API_STREAMING=true
#For Azure
#For Azure (you can delete it if you don't use Azure)
OPENAI_API_BASE=
OPENAI_API_VERSION=
AZURE_DEPLOYMENT_NAME=

138
.github/ISSUE_TEMPLATE/bug_report.yml vendored Normal file
View File

@@ -0,0 +1,138 @@
name: "🐛 Bug Report"
description: "Submit a bug report to help us improve"
title: "🐛 Bug Report: "
labels: ["type: bug"]
body:
- type: markdown
attributes:
value: We value your time and your efforts to submit this bug report is appreciated. 🙏
- type: textarea
id: description
validations:
required: true
attributes:
label: "📜 Description"
description: "A clear and concise description of what the bug is."
placeholder: "It bugs out when ..."
- type: textarea
id: steps-to-reproduce
validations:
required: true
attributes:
label: "👟 Reproduction steps"
description: "How do you trigger this bug? Please walk us through it step by step."
placeholder: "1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error"
- type: textarea
id: expected-behavior
validations:
required: true
attributes:
label: "👍 Expected behavior"
description: "What did you think should happen?"
placeholder: "It should ..."
- type: textarea
id: actual-behavior
validations:
required: true
attributes:
label: "👎 Actual Behavior with Screenshots"
description: "What did actually happen? Add screenshots, if applicable."
placeholder: "It actually ..."
- type: dropdown
id: operating-system
attributes:
label: "💻 Operating system"
description: "What OS is your app running on?"
options:
- Linux
- MacOS
- Windows
- Something else
validations:
required: true
- type: dropdown
id: browsers
attributes:
label: What browsers are you seeing the problem on?
multiple: true
options:
- Firefox
- Chrome
- Safari
- Microsoft Edge
- Something else
- type: dropdown
id: dev-environment
validations:
required: true
attributes:
label: "🤖 What development environment are you experiencing this bug on?"
options:
- Docker
- Local dev server
- type: textarea
id: env-vars
validations:
required: false
attributes:
label: "🔒 Did you set the correct environment variables in the right path? List the environment variable names (not values please!)"
description: "Please refer to the [Project setup instructions](https://github.com/arc53/DocsGPT#quickstart) if you are unsure."
placeholder: "It actually ..."
- type: textarea
id: additional-context
validations:
required: false
attributes:
label: "📃 Provide any additional context for the Bug."
description: "Add any other context about the problem here."
placeholder: "It actually ..."
- type: textarea
id: logs
validations:
required: false
attributes:
label: 📖 Relevant log output
description: Please copy and paste any relevant log output. This will be automatically formatted into code, so no need for backticks.
render: shell
- type: checkboxes
id: no-duplicate-issues
attributes:
label: "👀 Have you spent some time to check if this bug has been raised before?"
options:
- label: "I checked and didn't find similar issue"
required: true
- type: dropdown
id: willing-to-submit-pr
attributes:
label: 🔗 Are you willing to submit PR?
description: This is absolutely not required, but we are happy to guide you in the contribution process.
options: # Added options key
- "Yes, I am willing to submit a PR!"
- "No"
validations:
required: false
- type: checkboxes
id: terms
attributes:
label: 🧑‍⚖️ Code of Conduct
description: By submitting this issue, you agree to follow our [Code of Conduct](https://github.com/arc53/DocsGPT/blob/main/CODE_OF_CONDUCT.md)
options:
- label: I agree to follow this project's Code of Conduct
required: true

View File

@@ -0,0 +1,54 @@
name: 🚀 Feature
description: "Submit a proposal for a new feature"
title: "🚀 Feature: "
labels: [feature]
body:
- type: markdown
attributes:
value: We value your time and your efforts to submit this bug report is appreciated. 🙏
- type: textarea
id: feature-description
validations:
required: true
attributes:
label: "🔖 Feature description"
description: "A clear and concise description of what the feature is."
placeholder: "You should add ..."
- type: textarea
id: pitch
validations:
required: true
attributes:
label: "🎤 Why is this feature needed ?"
description: "Please explain why this feature should be implemented and how it would be used. Add examples, if applicable."
placeholder: "In my use-case, ..."
- type: textarea
id: solution
validations:
required: true
attributes:
label: "✌️ How do you aim to achieve this?"
description: "A clear and concise description of what you want to happen."
placeholder: "I want this feature to, ..."
- type: textarea
id: alternative
validations:
required: false
attributes:
label: "🔄️ Additional Information"
description: "A clear and concise description of any alternative solutions or additional solutions you've considered."
placeholder: "I tried, ..."
- type: checkboxes
id: no-duplicate-issues
attributes:
label: "👀 Have you spent some time to check if this feature request has been raised before?"
options:
- label: "I checked and didn't find similar issue"
required: true
- type: dropdown
id: willing-to-submit-pr
attributes:
label: Are you willing to submit PR?
description: This is absolutely not required, but we are happy to guide you in the contribution process.
options:
- "Yes I am willing to submit a PR!"

5
.github/PULL_REQUEST_TEMPLATE.md vendored Normal file
View File

@@ -0,0 +1,5 @@
- **What kind of change does this PR introduce?** (Bug fix, feature, docs update, ...)
- **Why was this change needed?** (You can also link to an open issue here)
- **Other information**:

5
.github/holopin.yml vendored Normal file
View File

@@ -0,0 +1,5 @@
organization: arc53
defaultSticker: clqmdf0ed34290glbvqh0kzxd
stickers:
- id: clqmdf0ed34290glbvqh0kzxd
alias: festive

23
.github/labeler.yml vendored Normal file
View File

@@ -0,0 +1,23 @@
repo:
- '*'
github:
- .github/**/*
application:
- application/**/*
docs:
- docs/**/*
extensions:
- extensions/**/*
frontend:
- frontend/**/*
scripts:
- scripts/**/*
tests:
- tests/**/*

15
.github/workflows/labeler.yml vendored Normal file
View File

@@ -0,0 +1,15 @@
# https://github.com/actions/labeler
name: Pull Request Labeler
on:
- pull_request_target
jobs:
triage:
permissions:
contents: read
pull-requests: write
runs-on: ubuntu-latest
steps:
- uses: actions/labeler@v4
with:
repo-token: "${{ secrets.GITHUB_TOKEN }}"
sync-labels: true

View File

@@ -1,15 +1,12 @@
name: Run python tests with pytest
on: [push, pull_request]
jobs:
build:
pytest_and_coverage:
name: Run tests and count coverage
runs-on: ubuntu-latest
strategy:
matrix:
python-version: ["3.9", "3.10", "3.11"]
steps:
- uses: actions/checkout@v3
- name: Set up Python ${{ matrix.python-version }}
@@ -19,9 +16,15 @@ jobs:
- name: Install dependencies
run: |
python -m pip install --upgrade pip
pip install pytest
pip install pytest pytest-cov
cd application
if [ -f requirements.txt ]; then pip install -r requirements.txt; fi
- name: Test with pytest
- name: Test with pytest and generate coverage report
run: |
python -m pytest
python -m pytest --cov=application --cov=scripts --cov=extensions --cov-report=xml
- name: Upload coverage reports to Codecov
if: github.event_name == 'pull_request' && matrix.python-version == '3.11'
uses: codecov/codecov-action@v3
env:
CODECOV_TOKEN: ${{ secrets.CODECOV_TOKEN }}

6
.gitignore vendored
View File

@@ -5,7 +5,7 @@ __pycache__/
# C extensions
*.so
*.next
# Distribution / packaging
.Python
build/
@@ -169,4 +169,6 @@ application/vectors/
**/yarn.lock
node_modules/
node_modules/
.vscode/settings.json
models/

Binary file not shown.

After

Width:  |  Height:  |  Size: 88 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 21 KiB

View File

@@ -2,58 +2,58 @@
## Our Pledge
We as members, contributors, and leaders pledge to make participation in our
community a harassment-free experience for everyone, regardless of age, body
We as members, contributors and leaders pledge to make participation in our
community, a harassment-free experience for everyone, regardless of age, body
size, visible or invisible disability, ethnicity, sex characteristics, gender
identity and expression, level of experience, education, socio-economic status,
nationality, personal appearance, race, religion, or sexual identity
nationality, personal appearance, race, religion or sexual identity
and orientation.
We pledge to act and interact in ways that contribute to an open, welcoming,
diverse, inclusive, and healthy community.
diverse, inclusive and a healthy community.
## Our Standards
Examples of behavior that contributes to a positive environment for our
Examples of behavior that contribute to a positive environment for our
community include:
* Demonstrating empathy and kindness toward other people
* Being respectful of differing opinions, viewpoints, and experiences
* Giving and gracefully accepting constructive feedback
* Accepting responsibility and apologizing to those affected by our mistakes,
and learning from the experience
* Focusing on what is best not just for us as individuals, but for the
overall community
## Demonstrating empathy and kindness towards other people
1. Being respectful and open to differing opinions, viewpoints, and experiences
2. Giving and gracefully accepting constructive feedback
3. Taking accountability and offering apologies to those who have been impacted by our errors,
while also gaining insights from the situation
4. Focusing on what is best not just for us as individuals but for the
community as a whole
Examples of unacceptable behavior include:
* The use of sexualized language or imagery, and sexual attention or
1. The use of sexualized language or imagery, and sexual attention or
advances of any kind
* Trolling, insulting or derogatory comments, and personal or political attacks
* Public or private harassment
* Publishing others' private information, such as a physical or email
2. Trolling, insulting or derogatory comments, and personal or political attacks
3. Public or private harassment
4. Publishing other's private information, such as a physical or email
address, without their explicit permission
* Other conduct which could reasonably be considered inappropriate in a
5. Other conduct which could reasonably be considered inappropriate in a
professional setting
## Enforcement Responsibilities
Community leaders are responsible for clarifying and enforcing our standards of
acceptable behavior and will take appropriate and fair corrective action in
response to any behavior that they deem inappropriate, threatening, offensive,
response to any behavior that they deem inappropriate, threatening, offensive
or harmful.
Community leaders have the right and responsibility to remove, edit, or reject
comments, commits, code, wiki edits, issues, and other contributions that are
not aligned to this Code of Conduct, and will communicate reasons for moderation
not aligned to this Code of Conduct and will communicate reasons for moderation
decisions when appropriate.
## Scope
This Code of Conduct applies within all community spaces, and also applies when
This Code of Conduct applies within all community spaces and also applies when
an individual is officially representing the community in public spaces.
Examples of representing our community include using an official e-mail address,
posting via an official social media account, or acting as an appointed
posting via an official social media account or acting as an appointed
representative at an online or offline event.
## Enforcement
@@ -63,29 +63,27 @@ reported to the community leaders responsible for enforcement at
contact@arc53.com.
All complaints will be reviewed and investigated promptly and fairly.
All community leaders are obligated to respect the privacy and security of the
All community leaders are obligated to be respectful towards the privacy and security of the
reporter of any incident.
## Enforcement Guidelines
Community leaders will follow these Community Impact Guidelines in determining
the consequences for any action they deem in violation of this Code of Conduct:
the consequences for any action that they deem in violation of this Code of Conduct:
### 1. Correction
* **Community Impact**: Use of inappropriate language or other behavior deemed
unprofessional or unwelcome in the community space.
**Community Impact**: Use of inappropriate language or other behavior deemed
unprofessional or unwelcome in the community.
**Consequence**: A private, written warning from community leaders, providing
* **Consequence**: A private, written warning from community leaders, providing
clarity around the nature of the violation and an explanation of why the
behavior was inappropriate. A public apology may be requested.
### 2. Warning
**Community Impact**: A violation through a single incident or series
* **Community Impact**: A violation through a single incident or series
of actions.
**Consequence**: A warning with consequences for continued behavior. No
* **Consequence**: A warning with consequences for continued behavior. No
interaction with the people involved, including unsolicited interaction with
those enforcing the Code of Conduct, for a specified period of time. This
includes avoiding interactions in community spaces as well as external channels
@@ -93,23 +91,21 @@ like social media. Violating these terms may lead to a temporary or
permanent ban.
### 3. Temporary Ban
**Community Impact**: A serious violation of community standards, including
* **Community Impact**: A serious violation of community standards, including
sustained inappropriate behavior.
**Consequence**: A temporary ban from any sort of interaction or public
* **Consequence**: A temporary ban from any sort of interaction or public
communication with the community for a specified period of time. No public or
private interaction with the people involved, including unsolicited interaction
with those enforcing the Code of Conduct, is allowed during this period.
Violating these terms may lead to a permanent ban.
### 4. Permanent Ban
* **Community Impact**: Demonstrating a pattern of violation of community
standards, including sustained inappropriate behavior,harassment of an
individual or aggression towards or disparagement of classes of individuals.
**Community Impact**: Demonstrating a pattern of violation of community
standards, including sustained inappropriate behavior, harassment of an
individual, or aggression toward or disparagement of classes of individuals.
**Consequence**: A permanent ban from any sort of public interaction within
* **Consequence**: A permanent ban from any sort of public interaction within
the community.
## Attribution

View File

@@ -1,38 +1,128 @@
# Welcome to DocsGPT Contributing guideline
# Welcome to DocsGPT Contributing Guidelines
Thank you for choosing this project to contribute to, we are all very grateful!
Thank you for choosing to contribute to DocsGPT! We are all very grateful!
# We accept different types of contributions
📣 Discussions - where you can start a new topic or answer some questions
📣 **Discussions** - Engage in conversations, start new topics, or help answer questions.
🐞 Issues - Is how we track tasks, sometimes its bugs that need fixing, sometimes its new features
🐞 **Issues** - This is where we keep track of tasks. It could be bugs,fixes or suggestions for new features.
🛠️ Pull requests - Is how you can suggest changes to our repository, to work on existing issue or to add new features
🛠️ **Pull requests** - Suggest changes to our repository, either by working on existing issues or adding new features.
📚 Wiki - where we have our documentation
📚 **Wiki** - This is where our documentation resides.
## 🐞 Issues and Pull requests
We value contributions to our issues in form of discussion or suggestion, we recommend that you check out existing issues and our [Roadmap](https://github.com/orgs/arc53/projects/2)
If you want to contribute by writing code there are few things that you should know before doing it:
We have frontend (React, Vite) and Backend (python)
### If you are looking to contribute to Frontend (⚛React, Vite):
Current frontend is being migrated from /application to /frontend with a new design, so please contribute to the new on. Check out this [Milestone](https://github.com/arc53/DocsGPT/milestone/1) and its issues also [Figma](https://www.figma.com/file/OXLtrl1EAy885to6S69554/DocsGPT?node-id=0%3A1&t=hjWVuxRg9yi5YkJ9-1)
Please try to follow guidelines
- We value contributions in the form of discussions or suggestions. We recommend taking a look at existing issues and our [roadmap](https://github.com/orgs/arc53/projects/2).
### If you are looking to contribute to Backend (🐍Python):
Check out our issues, and contribute to /application or /scripts (ignore old ingest_rst.py ingest_rst_sphinx.py files, they will be deprecated soon)
Currently we don't have any tests(which would be useful😉) but before submitting you PR make sure that after you ingested some test data its queryable
- If you're interested in contributing code, here are some important things to know:
### Workflow:
Create a fork, make changes on your forked repository, submit changes in a form of pull request
- We have a frontend built on React (Vite) and a backend in Python.
=======
Before creating issues, please check out how the latest version of our app looks and works by launching it via [Quickstart](https://github.com/arc53/DocsGPT#quickstart) the version on our live demo is slightly modified with login. Your issues should relate to the version that you can launch via [Quickstart](https://github.com/arc53/DocsGPT#quickstart).
## Questions / collaboration
Please join our [Discord](https://discord.gg/n5BX8dh8rU) don't hesitate, we are very friendly and welcoming to new contributors.
### 👨‍💻 If you're interested in contributing code, here are some important things to know:
# Thank you so much for considering to contribute to DocsGPT!🙏
Tech Stack Overview:
- 🌐 Frontend: Built with React (Vite) ⚛️,
- 🖥 Backend: Developed in Python 🐍
### 🌐 If you are looking to contribute to frontend (⚛React, Vite):
- The current frontend is being migrated from [`/application`](https://github.com/arc53/DocsGPT/tree/main/application) to [`/frontend`](https://github.com/arc53/DocsGPT/tree/main/frontend) with a new design, so please contribute to the new one.
- Check out this [milestone](https://github.com/arc53/DocsGPT/milestone/1) and its issues.
- The updated Figma design can be found [here](https://www.figma.com/file/OXLtrl1EAy885to6S69554/DocsGPT?node-id=0%3A1&t=hjWVuxRg9yi5YkJ9-1).
Please try to follow the guidelines.
### 🖥 If you are looking to contribute to Backend (🐍 Python):
- Review our issues and contribute to [`/application`](https://github.com/arc53/DocsGPT/tree/main/application) or [`/scripts`](https://github.com/arc53/DocsGPT/tree/main/scripts) (please disregard old [`ingest_rst.py`](https://github.com/arc53/DocsGPT/blob/main/scripts/old/ingest_rst.py) [`ingest_rst_sphinx.py`](https://github.com/arc53/DocsGPT/blob/main/scripts/old/ingest_rst_sphinx.py) files; they will be deprecated soon).
- All new code should be covered with unit tests ([pytest](https://github.com/pytest-dev/pytest)). Please find tests under [`/tests`](https://github.com/arc53/DocsGPT/tree/main/tests) folder.
- Before submitting your Pull Request, ensure it can be queried after ingesting some test data.
### Testing
To run unit tests from the root of the repository, execute:
```
python -m pytest
```
## Workflow 📈
Here's a step-by-step guide on how to contribute to DocsGPT:
1. **Fork the Repository:**
- Click the "Fork" button at the top-right of this repository to create your fork.
2. **Clone the Forked Repository:**
- Clone the repository using:
``` shell
git clone https://github.com/<your-github-username>/DocsGPT.git
```
3. **Keep your Fork in Sync:**
- Before you make any changes, make sure that your fork is in sync to avoid merge conflicts using:
```shell
git remote add upstream https://github.com/arc53/DocsGPT.git
git pull upstream main
```
4. **Create and Switch to a New Branch:**
- Create a new branch for your contribution using:
```shell
git checkout -b your-branch-name
```
5. **Make Changes:**
- Make the required changes in your branch.
6. **Add Changes to the Staging Area:**
- Add your changes to the staging area using:
```shell
git add .
```
7. **Commit Your Changes:**
- Commit your changes with a descriptive commit message using:
```shell
git commit -m "Your descriptive commit message"
```
8. **Push Your Changes to the Remote Repository:**
- Push your branch with changes to your fork on GitHub using:
```shell
git push origin your-branch-name
```
9. **Submit a Pull Request (PR):**
- Create a Pull Request from your branch to the main repository. Make sure to include a detailed description of your changes and reference any related issues.
10. **Collaborate:**
- Be responsive to comments and feedback on your PR.
- Make necessary updates as suggested.
- Once your PR is approved, it will be merged into the main repository.
11. **Testing:**
- Before submitting a Pull Request, ensure your code passes all unit tests.
- To run unit tests from the root of the repository, execute:
```shell
python -m pytest
```
*Note: You should run the unit test only after making the changes to the backend code.*
12. **Questions and Collaboration:**
- Feel free to join our Discord. We're very friendly and welcoming to new contributors, so don't hesitate to reach out.
Thank you for considering contributing to DocsGPT! 🙏
## Questions/collaboration
Feel free to join our [Discord](https://discord.gg/n5BX8dh8rU). We're very friendly and welcoming to new contributors, so don't hesitate to reach out.
# Thank you so much for considering to contribute DocsGPT!🙏

195
README.md
View File

@@ -7,130 +7,193 @@
</p>
<p align="left">
<strong>DocsGPT</strong> is a cutting-edge open-source solution that streamlines the process of finding information in project documentation. With its integration of the powerful <strong>GPT</strong> models, developers can easily ask questions about a project and receive accurate answers.
<strong><a href="https://docsgpt.arc53.com/">DocsGPT</a></strong> is a cutting-edge open-source solution that streamlines the process of finding information in the project documentation. With its integration of the powerful <strong>GPT</strong> models, developers can easily ask questions about a project and receive accurate answers.
Say goodbye to time-consuming manual searches, and let <strong>DocsGPT</strong> help you quickly find the information you need. Try it out and see how it revolutionizes your project documentation experience. Contribute to its development and be a part of the future of AI-powered assistance.
Say goodbye to time-consuming manual searches, and let <strong><a href="https://docsgpt.arc53.com/">DocsGPT</a></strong> help you quickly find the information you need. Try it out and see how it revolutionizes your project documentation experience. Contribute to its development and be a part of the future of AI-powered assistance.
</p>
<div align="center">
<a href="https://discord.gg/n5BX8dh8rU">![example1](https://img.shields.io/github/stars/arc53/docsgpt?style=social)</a>
<a href="https://discord.gg/n5BX8dh8rU">![example2](https://img.shields.io/github/forks/arc53/docsgpt?style=social)</a>
<a href="https://discord.gg/n5BX8dh8rU">![example3](https://img.shields.io/github/license/arc53/docsgpt)</a>
<a href="https://discord.gg/n5BX8dh8rU">![example3](https://img.shields.io/discord/1070046503302877216)</a>
<a href="https://github.com/arc53/DocsGPT">![link to main GitHub showing Stars number](https://img.shields.io/github/stars/arc53/docsgpt?style=social)</a>
<a href="https://github.com/arc53/DocsGPT">![link to main GitHub showing Forks number](https://img.shields.io/github/forks/arc53/docsgpt?style=social)</a>
<a href="https://github.com/arc53/DocsGPT/blob/main/LICENSE">![link to license file](https://img.shields.io/github/license/arc53/docsgpt)</a>
<a href="https://discord.gg/n5BX8dh8rU">![link to discord](https://img.shields.io/discord/1070046503302877216)</a>
<a href="https://twitter.com/docsgptai">![X (formerly Twitter) URL](https://img.shields.io/twitter/follow/docsgptai)</a>
</div>
### Production Support / Help for Companies:
We're eager to provide personalized assistance when deploying your DocsGPT to a live environment.
- [Book Demo :wave:](https://airtable.com/appdeaL0F1qV8Bl2C/shrrJF1Ll7btCJRbP)
- [Send Email :email:](mailto:contact@arc53.com?subject=DocsGPT%20support%2Fsolutions)
![video-example-of-docs-gpt](https://d3dg1063dc54p9.cloudfront.net/videos/demov3.gif)
## Roadmap
You can find our [Roadmap](https://github.com/orgs/arc53/projects/2) here, please don't hesitate contributing or creating issues, it helps us make DocsGPT better!
You can find our roadmap [here](https://github.com/orgs/arc53/projects/2). Please don't hesitate to contribute or create issues, it helps us improve DocsGPT!
## Our open source models optimised for DocsGPT:
## Our Open-Source Models Optimized for DocsGPT:
| Name | Base Model | Requirements (or similar) |
|-------------------|------------|----------------------------------------------------------|
| [Docsgpt-7b-falcon](https://huggingface.co/Arc53/docsgpt-7b-falcon) | Falcon-7b | 1xA10G gpu |
| [Docsgpt-14b](https://huggingface.co/Arc53/docsgpt-14b) | llama-2-14b | 2xA10 gpu's |
| [Docsgpt-40b](https://huggingface.co/Arc53/docsgpt-40b-falcon) | falcon-40b | 8xA10G gpu's |
If you don't have enough resources to run it you can use bitsnbytes to quantize
| Name | Base Model | Requirements (or similar) |
| --------------------------------------------------------------------- | ----------- | ------------------------- |
| [Docsgpt-7b-falcon](https://huggingface.co/Arc53/docsgpt-7b-falcon) | Falcon-7b | 1xA10G gpu |
| [Docsgpt-14b](https://huggingface.co/Arc53/docsgpt-14b) | llama-2-14b | 2xA10 gpu's |
| [Docsgpt-40b-falcon](https://huggingface.co/Arc53/docsgpt-40b-falcon) | falcon-40b | 8xA10G gpu's |
If you don't have enough resources to run it, you can use bitsnbytes to quantize.
## Features
![Group 9](https://user-images.githubusercontent.com/17906039/220427472-2644cff4-7666-46a5-819f-fc4a521f63c7.png)
![Main features of DocsGPT showcasing six main features](https://user-images.githubusercontent.com/17906039/220427472-2644cff4-7666-46a5-819f-fc4a521f63c7.png)
## Useful Links
## Useful links
[Live preview](https://docsgpt.arc53.com/)
[Join Our Discord](https://discord.gg/n5BX8dh8rU)
[Guides](https://github.com/arc53/docsgpt/wiki)
- :mag: :fire: [Live preview](https://docsgpt.arc53.com/)
[Interested in contributing?](https://github.com/arc53/DocsGPT/blob/main/CONTRIBUTING.md)
- :speech_balloon: :tada: [Join our Discord](https://discord.gg/n5BX8dh8rU)
[How to use any other documentation](https://github.com/arc53/docsgpt/wiki/How-to-train-on-other-documentation)
- :books: :sunglasses: [Guides](https://docs.docsgpt.co.uk/)
[How to host it locally (so all data will stay on-premises)](https://github.com/arc53/DocsGPT/wiki/How-to-use-different-LLM's#hosting-everything-locally)
- :couple: [Interested in contributing?](https://github.com/arc53/DocsGPT/blob/main/CONTRIBUTING.md)
- :file_folder: :rocket: [How to use any other documentation](https://docs.docsgpt.co.uk/Guides/How-to-train-on-other-documentation)
## Project structure
- Application - Flask app (main application)
- :house: :closed_lock_with_key: [How to host it locally (so all data will stay on-premises)](https://docs.docsgpt.co.uk/Guides/How-to-use-different-LLM)
- Extensions - Chrome extension
## Project Structure
- Scripts - Script that creates similarity search index and store for other libraries.
- Application - Flask app (main application).
- Frontend - Frontend uses Vite and React
- Extensions - Chrome extension.
- Scripts - Script that creates similarity search index for other libraries.
- Frontend - Frontend uses <a href="https://vitejs.dev/">Vite</a> and <a href="https://react.dev/">React</a>.
## QuickStart
Note: Make sure you have docker installed
> [!Note]
> Make sure you have [Docker](https://docs.docker.com/engine/install/) installed
1. Dowload and open this repository with `git clone https://github.com/arc53/DocsGPT.git`
2. Create an .env file in your root directory and set the env variable OPENAI_API_KEY with your openai api key and VITE_API_STREAMING to true or false, depending on if you want streaming answers or not
On Mac OS or Linux, write:
`./setup.sh`
It will install all the dependencies and allow you to download the local model, use OpenAI or use our LLM API.
Otherwise, refer to this Guide:
1. Download and open this repository with `git clone https://github.com/arc53/DocsGPT.git`
2. Create a `.env` file in your root directory and set the env variables and `VITE_API_STREAMING` to true or false, depending on whether you want streaming answers or not.
It should look like this inside:
```
OPENAI_API_KEY=Yourkey
LLM_NAME=[docsgpt or openai or others]
VITE_API_STREAMING=true
API_KEY=[if LLM_NAME is openai]
```
3. Run `./run-with-docker-compose.sh`
4. Navigate to http://localhost:5173/
To stop just run Ctrl + C
See optional environment variables in the [/.env-template](https://github.com/arc53/DocsGPT/blob/main/.env-template) and [/application/.env_sample](https://github.com/arc53/DocsGPT/blob/main/application/.env_sample) files.
## Development environments
3. Run [./run-with-docker-compose.sh](https://github.com/arc53/DocsGPT/blob/main/run-with-docker-compose.sh).
4. Navigate to http://localhost:5173/.
### Spin up mongo and redis
For development only 2 containers are used from docker-compose.yaml (by deleting all services except for redis and mongo).
To stop, just run `Ctrl + C`.
## Development Environments
### Spin up Mongo and Redis
For development, only two containers are used from [docker-compose.yaml](https://github.com/arc53/DocsGPT/blob/main/docker-compose.yaml) (by deleting all services except for Redis and Mongo).
See file [docker-compose-dev.yaml](./docker-compose-dev.yaml).
Run
```
docker compose -f docker-compose-dev.yaml build
docker compose -f docker-compose-dev.yaml up -d
```
### Run the backend
### Run the Backend
Make sure you have Python 3.10 or 3.11 installed.
> [!Note]
> Make sure you have Python 3.10 or 3.11 installed.
1. Export required environment variables or prepare a `.env` file in the `/application` folder:
- Copy [.env_sample](https://github.com/arc53/DocsGPT/blob/main/application/.env_sample) and create `.env`.
(check out [`application/core/settings.py`](application/core/settings.py) if you want to see more config options.)
2. (optional) Create a Python virtual environment:
You can follow the [Python official documentation](https://docs.python.org/3/tutorial/venv.html) for virtual environments.
a) On Mac OS and Linux
1. Export required environment variables
```commandline
export CELERY_BROKER_URL=redis://localhost:6379/0
export CELERY_RESULT_BACKEND=redis://localhost:6379/1
export MONGO_URI=mongodb://localhost:27017/docsgpt
```
2. Prepare .env file
Copy `.env_sample` and create `.env` with your OpenAI API token
3. (optional) Create a python virtual environment
```commandline
python -m venv venv
. venv/bin/activate
```
4. Change to `application/` subdir and install dependencies for the backend
b) On Windows
```commandline
python -m venv venv
venv/Scripts/activate
```
3. Download embedding model and save it in the `model/` folder:
You can use the script below, or download it manually from [here](https://d3dg1063dc54p9.cloudfront.net/models/embeddings/mpnet-base-v2.zip), unzip it and save it in the `model/` folder.
```commandline
wget https://d3dg1063dc54p9.cloudfront.net/models/embeddings/mpnet-base-v2.zip
unzip mpnet-base-v2.zip -d model
rm mpnet-base-v2.zip
4. Change to the `application/` subdir by the command `cd application/` and install dependencies for the backend:
```commandline
cd application/
pip install -r requirements.txt
```
5. Run the app `python wsgi.py`
6. Start worker with `celery -A app.celery worker -l INFO`
### Start frontend
Make sure you have Node version 16 or higher.
5. Run the app using `flask --app application/app.py run --host=0.0.0.0 --port=7091`.
6. Start worker with `celery -A application.app.celery worker -l INFO`.
1. Navigate to `/frontend` folder
2. Install dependencies
`npm install`
3. Run the app
`npm run dev`
### Start Frontend
> [!Note]
> Make sure you have Node version 16 or higher.
1. Navigate to the [/frontend](https://github.com/arc53/DocsGPT/tree/main/frontend) folder.
2. Install the required packages `husky` and `vite` (ignore if already installed).
Built with [🦜️🔗 LangChain](https://github.com/hwchase17/langchain)
```commandline
npm install husky -g
npm install vite -g
```
3. Install dependencies by running `npm install --include=dev`.
4. Run the app using `npm run dev`.
## Contributing
Please refer to the [CONTRIBUTING.md](CONTRIBUTING.md) file for information about how to get involved. We welcome issues, questions, and pull requests.
## Code Of Conduct
We as members, contributors, and leaders, pledge to make participation in our community a harassment-free experience for everyone, regardless of age, body size, visible or invisible disability, ethnicity, sex characteristics, gender identity and expression, level of experience, education, socio-economic status, nationality, personal appearance, race, religion, or sexual identity and orientation. Please refer to the [CODE_OF_CONDUCT.md](CODE_OF_CONDUCT.md) file for more information about contributing.
## Many Thanks To Our Contributors⚡
<a href="https://github.com/arc53/DocsGPT/graphs/contributors" alt="View Contributors">
<img src="https://contrib.rocks/image?repo=arc53/DocsGPT" alt="Contributors" />
</a>
## License
The source code license is [MIT](https://opensource.org/license/mit/), as described in the [LICENSE](LICENSE) file.
Built with [:bird: :link: LangChain](https://github.com/hwchase17/langchain)

View File

@@ -1,9 +1,8 @@
API_KEY=your_api_key
EMBEDDINGS_KEY=your_api_key
CELERY_BROKER_URL=redis://localhost:6379/0
CELERY_RESULT_BACKEND=redis://localhost:6379/1
MONGO_URI=mongodb://localhost:27017/docsgpt
API_URL=http://localhost:7091
FLASK_APP=application/app.py
FLASK_DEBUG=true
#For OPENAI on Azure
OPENAI_API_BASE=

View File

@@ -1,19 +1,25 @@
FROM python:3.10-slim-bullseye as builder
FROM python:3.11-slim-bullseye as builder
# Tiktoken requires Rust toolchain, so build it in a separate stage
RUN apt-get update && apt-get install -y gcc curl
RUN curl https://sh.rustup.rs -sSf | sh -s -- -y && apt-get install --reinstall libc6-dev -y
ENV PATH="/root/.cargo/bin:${PATH}"
RUN pip install --upgrade pip && pip install tiktoken==0.3.3
RUN pip install --upgrade pip && pip install tiktoken==0.5.2
COPY requirements.txt .
RUN pip install -r requirements.txt
RUN apt-get install -y wget unzip
RUN wget https://d3dg1063dc54p9.cloudfront.net/models/embeddings/mpnet-base-v2.zip
RUN unzip mpnet-base-v2.zip -d model
RUN rm mpnet-base-v2.zip
FROM python:3.10-slim-bullseye
FROM python:3.11-slim-bullseye
# Copy pre-built packages and binaries from builder stage
COPY --from=builder /usr/local/ /usr/local/
WORKDIR /app
COPY --from=builder /model /app/model
COPY . /app/application
ENV FLASK_APP=app.py
ENV FLASK_DEBUG=true

View File

View File

View File

@@ -0,0 +1,374 @@
import asyncio
import os
from flask import Blueprint, request, Response
import json
import datetime
import logging
import traceback
from pymongo import MongoClient
from bson.objectid import ObjectId
from transformers import GPT2TokenizerFast
from application.core.settings import settings
from application.vectorstore.vector_creator import VectorCreator
from application.llm.llm_creator import LLMCreator
from application.error import bad_request
logger = logging.getLogger(__name__)
mongo = MongoClient(settings.MONGO_URI)
db = mongo["docsgpt"]
conversations_collection = db["conversations"]
vectors_collection = db["vectors"]
prompts_collection = db["prompts"]
answer = Blueprint('answer', __name__)
if settings.LLM_NAME == "gpt4":
gpt_model = 'gpt-4'
elif settings.LLM_NAME == "anthropic":
gpt_model = 'claude-2'
else:
gpt_model = 'gpt-3.5-turbo'
# load the prompts
current_dir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
with open(os.path.join(current_dir, "prompts", "chat_combine_default.txt"), "r") as f:
chat_combine_template = f.read()
with open(os.path.join(current_dir, "prompts", "chat_reduce_prompt.txt"), "r") as f:
chat_reduce_template = f.read()
with open(os.path.join(current_dir, "prompts", "chat_combine_creative.txt"), "r") as f:
chat_combine_creative = f.read()
with open(os.path.join(current_dir, "prompts", "chat_combine_strict.txt"), "r") as f:
chat_combine_strict = f.read()
api_key_set = settings.API_KEY is not None
embeddings_key_set = settings.EMBEDDINGS_KEY is not None
async def async_generate(chain, question, chat_history):
result = await chain.arun({"question": question, "chat_history": chat_history})
return result
def count_tokens(string):
tokenizer = GPT2TokenizerFast.from_pretrained('gpt2')
return len(tokenizer(string)['input_ids'])
def run_async_chain(chain, question, chat_history):
loop = asyncio.new_event_loop()
asyncio.set_event_loop(loop)
result = {}
try:
answer = loop.run_until_complete(async_generate(chain, question, chat_history))
finally:
loop.close()
result["answer"] = answer
return result
def get_vectorstore(data):
if "active_docs" in data:
if data["active_docs"].split("/")[0] == "default":
vectorstore = ""
elif data["active_docs"].split("/")[0] == "local":
vectorstore = "indexes/" + data["active_docs"]
else:
vectorstore = "vectors/" + data["active_docs"]
if data["active_docs"] == "default":
vectorstore = ""
else:
vectorstore = ""
vectorstore = os.path.join("application", vectorstore)
return vectorstore
def is_azure_configured():
return settings.OPENAI_API_BASE and settings.OPENAI_API_VERSION and settings.AZURE_DEPLOYMENT_NAME
def complete_stream(question, docsearch, chat_history, api_key, prompt_id, conversation_id):
llm = LLMCreator.create_llm(settings.LLM_NAME, api_key=api_key)
if prompt_id == 'default':
prompt = chat_combine_template
elif prompt_id == 'creative':
prompt = chat_combine_creative
elif prompt_id == 'strict':
prompt = chat_combine_strict
else:
prompt = prompts_collection.find_one({"_id": ObjectId(prompt_id)})["content"]
docs = docsearch.search(question, k=2)
if settings.LLM_NAME == "llama.cpp":
docs = [docs[0]]
# join all page_content together with a newline
docs_together = "\n".join([doc.page_content for doc in docs])
p_chat_combine = prompt.replace("{summaries}", docs_together)
messages_combine = [{"role": "system", "content": p_chat_combine}]
source_log_docs = []
for doc in docs:
if doc.metadata:
source_log_docs.append({"title": doc.metadata['title'].split('/')[-1], "text": doc.page_content})
else:
source_log_docs.append({"title": doc.page_content, "text": doc.page_content})
if len(chat_history) > 1:
tokens_current_history = 0
# count tokens in history
chat_history.reverse()
for i in chat_history:
if "prompt" in i and "response" in i:
tokens_batch = count_tokens(i["prompt"]) + count_tokens(i["response"])
if tokens_current_history + tokens_batch < settings.TOKENS_MAX_HISTORY:
tokens_current_history += tokens_batch
messages_combine.append({"role": "user", "content": i["prompt"]})
messages_combine.append({"role": "system", "content": i["response"]})
messages_combine.append({"role": "user", "content": question})
response_full = ""
completion = llm.gen_stream(model=gpt_model, engine=settings.AZURE_DEPLOYMENT_NAME,
messages=messages_combine)
for line in completion:
data = json.dumps({"answer": str(line)})
response_full += str(line)
yield f"data: {data}\n\n"
# save conversation to database
if conversation_id is not None:
conversations_collection.update_one(
{"_id": ObjectId(conversation_id)},
{"$push": {"queries": {"prompt": question, "response": response_full, "sources": source_log_docs}}},
)
else:
# create new conversation
# generate summary
messages_summary = [{"role": "assistant", "content": "Summarise following conversation in no more than 3 "
"words, respond ONLY with the summary, use the same "
"language as the system \n\nUser: " + question + "\n\n" +
"AI: " +
response_full},
{"role": "user", "content": "Summarise following conversation in no more than 3 words, "
"respond ONLY with the summary, use the same language as the "
"system"}]
completion = llm.gen(model=gpt_model, engine=settings.AZURE_DEPLOYMENT_NAME,
messages=messages_summary, max_tokens=30)
conversation_id = conversations_collection.insert_one(
{"user": "local",
"date": datetime.datetime.utcnow(),
"name": completion,
"queries": [{"prompt": question, "response": response_full, "sources": source_log_docs}]}
).inserted_id
# send data.type = "end" to indicate that the stream has ended as json
data = json.dumps({"type": "id", "id": str(conversation_id)})
yield f"data: {data}\n\n"
data = json.dumps({"type": "end"})
yield f"data: {data}\n\n"
@answer.route("/stream", methods=["POST"])
def stream():
data = request.get_json()
# get parameter from url question
question = data["question"]
history = data["history"]
# history to json object from string
history = json.loads(history)
conversation_id = data["conversation_id"]
if 'prompt_id' in data:
prompt_id = data["prompt_id"]
else:
prompt_id = 'default'
# check if active_docs is set
if not api_key_set:
api_key = data["api_key"]
else:
api_key = settings.API_KEY
if not embeddings_key_set:
embeddings_key = data["embeddings_key"]
else:
embeddings_key = settings.EMBEDDINGS_KEY
if "active_docs" in data:
vectorstore = get_vectorstore({"active_docs": data["active_docs"]})
else:
vectorstore = ""
docsearch = VectorCreator.create_vectorstore(settings.VECTOR_STORE, vectorstore, embeddings_key)
return Response(
complete_stream(question, docsearch,
chat_history=history, api_key=api_key,
prompt_id=prompt_id,
conversation_id=conversation_id), mimetype="text/event-stream"
)
@answer.route("/api/answer", methods=["POST"])
def api_answer():
data = request.get_json()
question = data["question"]
history = data["history"]
if "conversation_id" not in data:
conversation_id = None
else:
conversation_id = data["conversation_id"]
print("-" * 5)
if not api_key_set:
api_key = data["api_key"]
else:
api_key = settings.API_KEY
if not embeddings_key_set:
embeddings_key = data["embeddings_key"]
else:
embeddings_key = settings.EMBEDDINGS_KEY
if 'prompt_id' in data:
prompt_id = data["prompt_id"]
else:
prompt_id = 'default'
if prompt_id == 'default':
prompt = chat_combine_template
elif prompt_id == 'creative':
prompt = chat_combine_creative
elif prompt_id == 'strict':
prompt = chat_combine_strict
else:
prompt = prompts_collection.find_one({"_id": ObjectId(prompt_id)})["content"]
# use try and except to check for exception
try:
# check if the vectorstore is set
vectorstore = get_vectorstore(data)
# loading the index and the store and the prompt template
# Note if you have used other embeddings than OpenAI, you need to change the embeddings
docsearch = VectorCreator.create_vectorstore(settings.VECTOR_STORE, vectorstore, embeddings_key)
llm = LLMCreator.create_llm(settings.LLM_NAME, api_key=api_key)
docs = docsearch.search(question, k=2)
# join all page_content together with a newline
docs_together = "\n".join([doc.page_content for doc in docs])
p_chat_combine = prompt.replace("{summaries}", docs_together)
messages_combine = [{"role": "system", "content": p_chat_combine}]
source_log_docs = []
for doc in docs:
if doc.metadata:
source_log_docs.append({"title": doc.metadata['title'].split('/')[-1], "text": doc.page_content})
else:
source_log_docs.append({"title": doc.page_content, "text": doc.page_content})
# join all page_content together with a newline
if len(history) > 1:
tokens_current_history = 0
# count tokens in history
history.reverse()
for i in history:
if "prompt" in i and "response" in i:
tokens_batch = count_tokens(i["prompt"]) + count_tokens(i["response"])
if tokens_current_history + tokens_batch < settings.TOKENS_MAX_HISTORY:
tokens_current_history += tokens_batch
messages_combine.append({"role": "user", "content": i["prompt"]})
messages_combine.append({"role": "system", "content": i["response"]})
messages_combine.append({"role": "user", "content": question})
completion = llm.gen(model=gpt_model, engine=settings.AZURE_DEPLOYMENT_NAME,
messages=messages_combine)
result = {"answer": completion, "sources": source_log_docs}
logger.debug(result)
# generate conversationId
if conversation_id is not None:
conversations_collection.update_one(
{"_id": ObjectId(conversation_id)},
{"$push": {"queries": {"prompt": question,
"response": result["answer"], "sources": result['sources']}}},
)
else:
# create new conversation
# generate summary
messages_summary = [
{"role": "assistant", "content": "Summarise following conversation in no more than 3 words, "
"respond ONLY with the summary, use the same language as the system \n\n"
"User: " + question + "\n\n" + "AI: " + result["answer"]},
{"role": "user", "content": "Summarise following conversation in no more than 3 words, "
"respond ONLY with the summary, use the same language as the system"}
]
completion = llm.gen(
model=gpt_model,
engine=settings.AZURE_DEPLOYMENT_NAME,
messages=messages_summary,
max_tokens=30
)
conversation_id = conversations_collection.insert_one(
{"user": "local",
"date": datetime.datetime.utcnow(),
"name": completion,
"queries": [{"prompt": question, "response": result["answer"], "sources": source_log_docs}]}
).inserted_id
result["conversation_id"] = str(conversation_id)
# mock result
# result = {
# "answer": "The answer is 42",
# "sources": ["https://en.wikipedia.org/wiki/42_(number)", "https://en.wikipedia.org/wiki/42_(number)"]
# }
return result
except Exception as e:
# print whole traceback
traceback.print_exc()
print(str(e))
return bad_request(500, str(e))
@answer.route("/api/search", methods=["POST"])
def api_search():
data = request.get_json()
# get parameter from url question
question = data["question"]
if not embeddings_key_set:
if "embeddings_key" in data:
embeddings_key = data["embeddings_key"]
else:
embeddings_key = settings.EMBEDDINGS_KEY
else:
embeddings_key = settings.EMBEDDINGS_KEY
if "active_docs" in data:
vectorstore = get_vectorstore({"active_docs": data["active_docs"]})
else:
vectorstore = ""
docsearch = VectorCreator.create_vectorstore(settings.VECTOR_STORE, vectorstore, embeddings_key)
docs = docsearch.search(question, k=2)
source_log_docs = []
for doc in docs:
if doc.metadata:
source_log_docs.append({"title": doc.metadata['title'].split('/')[-1], "text": doc.page_content})
else:
source_log_docs.append({"title": doc.page_content, "text": doc.page_content})
#yield f"data:{data}\n\n"
return source_log_docs

View File

View File

@@ -0,0 +1,69 @@
import os
import datetime
from flask import Blueprint, request, send_from_directory
from pymongo import MongoClient
from werkzeug.utils import secure_filename
from application.core.settings import settings
mongo = MongoClient(settings.MONGO_URI)
db = mongo["docsgpt"]
conversations_collection = db["conversations"]
vectors_collection = db["vectors"]
current_dir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
internal = Blueprint('internal', __name__)
@internal.route("/api/download", methods=["get"])
def download_file():
user = secure_filename(request.args.get("user"))
job_name = secure_filename(request.args.get("name"))
filename = secure_filename(request.args.get("file"))
save_dir = os.path.join(current_dir, settings.UPLOAD_FOLDER, user, job_name)
return send_from_directory(save_dir, filename, as_attachment=True)
@internal.route("/api/upload_index", methods=["POST"])
def upload_index_files():
"""Upload two files(index.faiss, index.pkl) to the user's folder."""
if "user" not in request.form:
return {"status": "no user"}
user = secure_filename(request.form["user"])
if "name" not in request.form:
return {"status": "no name"}
job_name = secure_filename(request.form["name"])
save_dir = os.path.join(current_dir, "indexes", user, job_name)
if settings.VECTOR_STORE == "faiss":
if "file_faiss" not in request.files:
print("No file part")
return {"status": "no file"}
file_faiss = request.files["file_faiss"]
if file_faiss.filename == "":
return {"status": "no file name"}
if "file_pkl" not in request.files:
print("No file part")
return {"status": "no file"}
file_pkl = request.files["file_pkl"]
if file_pkl.filename == "":
return {"status": "no file name"}
# saves index files
if not os.path.exists(save_dir):
os.makedirs(save_dir)
file_faiss.save(os.path.join(save_dir, "index.faiss"))
file_pkl.save(os.path.join(save_dir, "index.pkl"))
# create entry in vectors_collection
vectors_collection.insert_one(
{
"user": user,
"name": job_name,
"language": job_name,
"location": save_dir,
"date": datetime.datetime.now().strftime("%d/%m/%Y %H:%M:%S"),
"model": settings.EMBEDDINGS_NAME,
"type": "local",
}
)
return {"status": "ok"}

View File

View File

@@ -0,0 +1,321 @@
import os
from flask import Blueprint, request, jsonify
import requests
from pymongo import MongoClient
from bson.objectid import ObjectId
from werkzeug.utils import secure_filename
from application.api.user.tasks import ingest
from application.core.settings import settings
from application.vectorstore.vector_creator import VectorCreator
mongo = MongoClient(settings.MONGO_URI)
db = mongo["docsgpt"]
conversations_collection = db["conversations"]
vectors_collection = db["vectors"]
prompts_collection = db["prompts"]
feedback_collection = db["feedback"]
user = Blueprint('user', __name__)
current_dir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
@user.route("/api/delete_conversation", methods=["POST"])
def delete_conversation():
# deletes a conversation from the database
conversation_id = request.args.get("id")
# write to mongodb
conversations_collection.delete_one(
{
"_id": ObjectId(conversation_id),
}
)
return {"status": "ok"}
@user.route("/api/get_conversations", methods=["get"])
def get_conversations():
# provides a list of conversations
conversations = conversations_collection.find().sort("date", -1)
list_conversations = []
for conversation in conversations:
list_conversations.append({"id": str(conversation["_id"]), "name": conversation["name"]})
#list_conversations = [{"id": "default", "name": "default"}, {"id": "jeff", "name": "jeff"}]
return jsonify(list_conversations)
@user.route("/api/get_single_conversation", methods=["get"])
def get_single_conversation():
# provides data for a conversation
conversation_id = request.args.get("id")
conversation = conversations_collection.find_one({"_id": ObjectId(conversation_id)})
return jsonify(conversation['queries'])
@user.route("/api/update_conversation_name", methods=["POST"])
def update_conversation_name():
# update data for a conversation
data = request.get_json()
id = data["id"]
name = data["name"]
conversations_collection.update_one({"_id": ObjectId(id)},{"$set":{"name":name}})
return {"status": "ok"}
@user.route("/api/feedback", methods=["POST"])
def api_feedback():
data = request.get_json()
question = data["question"]
answer = data["answer"]
feedback = data["feedback"]
feedback_collection.insert_one(
{
"question": question,
"answer": answer,
"feedback": feedback,
}
)
return {"status": "ok"}
@user.route("/api/delete_by_ids", methods=["get"])
def delete_by_ids():
"""Delete by ID. These are the IDs in the vectorstore"""
ids = request.args.get("path")
if not ids:
return {"status": "error"}
if settings.VECTOR_STORE == "faiss":
result = vectors_collection.delete_index(ids=ids)
if result:
return {"status": "ok"}
return {"status": "error"}
@user.route("/api/delete_old", methods=["get"])
def delete_old():
"""Delete old indexes."""
import shutil
path = request.args.get("path")
dirs = path.split("/")
dirs_clean = []
for i in range(0, len(dirs)):
dirs_clean.append(secure_filename(dirs[i]))
# check that path strats with indexes or vectors
if dirs_clean[0] not in ["indexes", "vectors"]:
return {"status": "error"}
path_clean = "/".join(dirs_clean)
vectors_collection.delete_one({"name": dirs_clean[-1], 'user': dirs_clean[-2]})
if settings.VECTOR_STORE == "faiss":
try:
shutil.rmtree(os.path.join(current_dir, path_clean))
except FileNotFoundError:
pass
else:
vetorstore = VectorCreator.create_vectorstore(
settings.VECTOR_STORE, path=os.path.join(current_dir, path_clean)
)
vetorstore.delete_index()
return {"status": "ok"}
@user.route("/api/upload", methods=["POST"])
def upload_file():
"""Upload a file to get vectorized and indexed."""
if "user" not in request.form:
return {"status": "no user"}
user = secure_filename(request.form["user"])
if "name" not in request.form:
return {"status": "no name"}
job_name = secure_filename(request.form["name"])
# check if the post request has the file part
if "file" not in request.files:
print("No file part")
return {"status": "no file"}
file = request.files["file"]
if file.filename == "":
return {"status": "no file name"}
if file:
filename = secure_filename(file.filename)
# save dir
save_dir = os.path.join(current_dir, settings.UPLOAD_FOLDER, user, job_name)
# create dir if not exists
if not os.path.exists(save_dir):
os.makedirs(save_dir)
file.save(os.path.join(save_dir, filename))
task = ingest.delay(settings.UPLOAD_FOLDER, [".rst", ".md", ".pdf", ".txt", ".docx",
".csv", ".epub", ".html", ".mdx"],
job_name, filename, user)
# task id
task_id = task.id
return {"status": "ok", "task_id": task_id}
else:
return {"status": "error"}
@user.route("/api/task_status", methods=["GET"])
def task_status():
"""Get celery job status."""
task_id = request.args.get("task_id")
from application.celery import celery
task = celery.AsyncResult(task_id)
task_meta = task.info
return {"status": task.status, "result": task_meta}
@user.route("/api/combine", methods=["GET"])
def combined_json():
user = "local"
"""Provide json file with combined available indexes."""
# get json from https://d3dg1063dc54p9.cloudfront.net/combined.json
data = [
{
"name": "default",
"language": "default",
"version": "",
"description": "default",
"fullName": "default",
"date": "default",
"docLink": "default",
"model": settings.EMBEDDINGS_NAME,
"location": "remote",
}
]
# structure: name, language, version, description, fullName, date, docLink
# append data from vectors_collection
for index in vectors_collection.find({"user": user}):
data.append(
{
"name": index["name"],
"language": index["language"],
"version": "",
"description": index["name"],
"fullName": index["name"],
"date": index["date"],
"docLink": index["location"],
"model": settings.EMBEDDINGS_NAME,
"location": "local",
}
)
if settings.VECTOR_STORE == "faiss":
data_remote = requests.get("https://d3dg1063dc54p9.cloudfront.net/combined.json").json()
for index in data_remote:
index["location"] = "remote"
data.append(index)
return jsonify(data)
@user.route("/api/docs_check", methods=["POST"])
def check_docs():
# check if docs exist in a vectorstore folder
data = request.get_json()
# split docs on / and take first part
if data["docs"].split("/")[0] == "local":
return {"status": "exists"}
vectorstore = "vectors/" + data["docs"]
base_path = "https://raw.githubusercontent.com/arc53/DocsHUB/main/"
if os.path.exists(vectorstore) or data["docs"] == "default":
return {"status": "exists"}
else:
r = requests.get(base_path + vectorstore + "index.faiss")
if r.status_code != 200:
return {"status": "null"}
else:
if not os.path.exists(vectorstore):
os.makedirs(vectorstore)
with open(vectorstore + "index.faiss", "wb") as f:
f.write(r.content)
# download the store
r = requests.get(base_path + vectorstore + "index.pkl")
with open(vectorstore + "index.pkl", "wb") as f:
f.write(r.content)
return {"status": "loaded"}
@user.route("/api/create_prompt", methods=["POST"])
def create_prompt():
data = request.get_json()
content = data["content"]
name = data["name"]
if name == "":
return {"status": "error"}
user = "local"
resp = prompts_collection.insert_one(
{
"name": name,
"content": content,
"user": user,
}
)
new_id = str(resp.inserted_id)
return {"id": new_id}
@user.route("/api/get_prompts", methods=["GET"])
def get_prompts():
user = "local"
prompts = prompts_collection.find({"user": user})
list_prompts = []
list_prompts.append({"id": "default", "name": "default", "type": "public"})
list_prompts.append({"id": "creative", "name": "creative", "type": "public"})
list_prompts.append({"id": "strict", "name": "strict", "type": "public"})
for prompt in prompts:
list_prompts.append({"id": str(prompt["_id"]), "name": prompt["name"], "type": "private"})
return jsonify(list_prompts)
@user.route("/api/get_single_prompt", methods=["GET"])
def get_single_prompt():
prompt_id = request.args.get("id")
if prompt_id == 'default':
with open(os.path.join(current_dir, "prompts", "chat_combine_default.txt"), "r") as f:
chat_combine_template = f.read()
return jsonify({"content": chat_combine_template})
elif prompt_id == 'creative':
with open(os.path.join(current_dir, "prompts", "chat_combine_creative.txt"), "r") as f:
chat_reduce_creative = f.read()
return jsonify({"content": chat_reduce_creative})
elif prompt_id == 'strict':
with open(os.path.join(current_dir, "prompts", "chat_combine_strict.txt"), "r") as f:
chat_reduce_strict = f.read()
return jsonify({"content": chat_reduce_strict})
prompt = prompts_collection.find_one({"_id": ObjectId(prompt_id)})
return jsonify({"content": prompt["content"]})
@user.route("/api/delete_prompt", methods=["POST"])
def delete_prompt():
data = request.get_json()
id = data["id"]
prompts_collection.delete_one(
{
"_id": ObjectId(id),
}
)
return {"status": "ok"}
@user.route("/api/update_prompt", methods=["POST"])
def update_prompt_name():
data = request.get_json()
id = data["id"]
name = data["name"]
content = data["content"]
# check if name is null
if name == "":
return {"status": "error"}
prompts_collection.update_one({"_id": ObjectId(id)},{"$set":{"name":name, "content": content}})
return {"status": "ok"}

View File

@@ -0,0 +1,7 @@
from application.worker import ingest_worker
from application.celery import celery
@celery.task(bind=True)
def ingest(self, directory, formats, name_job, filename, user):
resp = ingest_worker(self, directory, formats, name_job, filename, user)
return resp

View File

@@ -1,719 +1,44 @@
import asyncio
import datetime
import http.client
import json
import logging
import os
import platform
import traceback
import dotenv
import openai
import requests
from celery import Celery
from celery.result import AsyncResult
from flask import Flask, request, render_template, send_from_directory, jsonify, Response
from langchain import FAISS
from langchain import VectorDBQA, Cohere, OpenAI
from langchain.chains import LLMChain, ConversationalRetrievalChain
from langchain.chains.conversational_retrieval.prompts import CONDENSE_QUESTION_PROMPT
from langchain.chains.question_answering import load_qa_chain
from langchain.chat_models import ChatOpenAI, AzureChatOpenAI
from langchain.embeddings import (
OpenAIEmbeddings,
HuggingFaceHubEmbeddings,
CohereEmbeddings,
HuggingFaceInstructEmbeddings,
)
from langchain.prompts import PromptTemplate
from langchain.prompts.chat import (
ChatPromptTemplate,
SystemMessagePromptTemplate,
HumanMessagePromptTemplate,
AIMessagePromptTemplate,
)
from langchain.schema import HumanMessage, AIMessage
from pymongo import MongoClient
from werkzeug.utils import secure_filename
from application.celery import celery
from flask import Flask, request, redirect
from application.core.settings import settings
from application.error import bad_request
from application.worker import ingest_worker
from bson.objectid import ObjectId
# os.environ["LANGCHAIN_HANDLER"] = "langchain"
logger = logging.getLogger(__name__)
if settings.LLM_NAME == "gpt4":
gpt_model = 'gpt-4'
else:
gpt_model = 'gpt-3.5-turbo'
if settings.SELF_HOSTED_MODEL:
from langchain.llms import HuggingFacePipeline
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
model_id = settings.LLM_NAME # hf model id (Arc53/docsgpt-7b-falcon, Arc53/docsgpt-14b)
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)
pipe = pipeline(
"text-generation", model=model,
tokenizer=tokenizer, max_new_tokens=2000,
device_map="auto", eos_token_id=tokenizer.eos_token_id
)
hf = HuggingFacePipeline(pipeline=pipe)
# Redirect PosixPath to WindowsPath on Windows
from application.api.user.routes import user
from application.api.answer.routes import answer
from application.api.internal.routes import internal
if platform.system() == "Windows":
import pathlib
temp = pathlib.PosixPath
pathlib.PosixPath = pathlib.WindowsPath
# loading the .env file
dotenv.load_dotenv()
# load the prompts
current_dir = os.path.dirname(os.path.abspath(__file__))
with open(os.path.join(current_dir, "prompts", "combine_prompt.txt"), "r") as f:
template = f.read()
with open(os.path.join(current_dir, "prompts", "combine_prompt_hist.txt"), "r") as f:
template_hist = f.read()
with open(os.path.join(current_dir, "prompts", "question_prompt.txt"), "r") as f:
template_quest = f.read()
with open(os.path.join(current_dir, "prompts", "chat_combine_prompt.txt"), "r") as f:
chat_combine_template = f.read()
with open(os.path.join(current_dir, "prompts", "chat_reduce_prompt.txt"), "r") as f:
chat_reduce_template = f.read()
api_key_set = settings.API_KEY is not None
embeddings_key_set = settings.EMBEDDINGS_KEY is not None
app = Flask(__name__)
app.config["UPLOAD_FOLDER"] = UPLOAD_FOLDER = "inputs"
app.config["CELERY_BROKER_URL"] = settings.CELERY_BROKER_URL
app.config["CELERY_RESULT_BACKEND"] = settings.CELERY_RESULT_BACKEND
app.config["MONGO_URI"] = settings.MONGO_URI
celery = Celery()
app.register_blueprint(user)
app.register_blueprint(answer)
app.register_blueprint(internal)
app.config.update(
UPLOAD_FOLDER="inputs",
CELERY_BROKER_URL=settings.CELERY_BROKER_URL,
CELERY_RESULT_BACKEND=settings.CELERY_RESULT_BACKEND,
MONGO_URI=settings.MONGO_URI
)
celery.config_from_object("application.celeryconfig")
mongo = MongoClient(app.config["MONGO_URI"])
db = mongo["docsgpt"]
vectors_collection = db["vectors"]
conversations_collection = db["conversations"]
async def async_generate(chain, question, chat_history):
result = await chain.arun({"question": question, "chat_history": chat_history})
return result
def run_async_chain(chain, question, chat_history):
loop = asyncio.new_event_loop()
asyncio.set_event_loop(loop)
result = {}
try:
answer = loop.run_until_complete(async_generate(chain, question, chat_history))
finally:
loop.close()
result["answer"] = answer
return result
def get_vectorstore(data):
if "active_docs" in data:
if data["active_docs"].split("/")[0] == "local":
if data["active_docs"].split("/")[1] == "default":
vectorstore = ""
else:
vectorstore = "indexes/" + data["active_docs"]
else:
vectorstore = "vectors/" + data["active_docs"]
if data["active_docs"] == "default":
vectorstore = ""
else:
vectorstore = ""
vectorstore = os.path.join("application", vectorstore)
return vectorstore
def get_docsearch(vectorstore, embeddings_key):
if settings.EMBEDDINGS_NAME == "openai_text-embedding-ada-002":
if is_azure_configured():
os.environ["OPENAI_API_TYPE"] = "azure"
openai_embeddings = OpenAIEmbeddings(model=settings.AZURE_EMBEDDINGS_DEPLOYMENT_NAME)
else:
openai_embeddings = OpenAIEmbeddings(openai_api_key=embeddings_key)
docsearch = FAISS.load_local(vectorstore, openai_embeddings)
elif settings.EMBEDDINGS_NAME == "huggingface_sentence-transformers/all-mpnet-base-v2":
docsearch = FAISS.load_local(vectorstore, HuggingFaceHubEmbeddings())
elif settings.EMBEDDINGS_NAME == "huggingface_hkunlp/instructor-large":
docsearch = FAISS.load_local(vectorstore, HuggingFaceInstructEmbeddings())
elif settings.EMBEDDINGS_NAME == "cohere_medium":
docsearch = FAISS.load_local(vectorstore, CohereEmbeddings(cohere_api_key=embeddings_key))
return docsearch
@celery.task(bind=True)
def ingest(self, directory, formats, name_job, filename, user):
resp = ingest_worker(self, directory, formats, name_job, filename, user)
return resp
@app.route("/")
def home():
return render_template(
"index.html", api_key_set=api_key_set, llm_choice=settings.LLM_NAME, embeddings_choice=settings.EMBEDDINGS_NAME
)
def complete_stream(question, docsearch, chat_history, api_key, conversation_id):
openai.api_key = api_key
if is_azure_configured():
logger.debug("in Azure")
openai.api_type = "azure"
openai.api_version = settings.OPENAI_API_VERSION
openai.api_base = settings.OPENAI_API_BASE
llm = AzureChatOpenAI(
openai_api_key=api_key,
openai_api_base=settings.OPENAI_API_BASE,
openai_api_version=settings.OPENAI_API_VERSION,
deployment_name=settings.AZURE_DEPLOYMENT_NAME,
)
if request.remote_addr in ('0.0.0.0', '127.0.0.1', 'localhost', '172.18.0.1'):
return redirect('http://localhost:5173')
else:
logger.debug("plain OpenAI")
llm = ChatOpenAI(openai_api_key=api_key)
docs = docsearch.similarity_search(question, k=2)
# join all page_content together with a newline
docs_together = "\n".join([doc.page_content for doc in docs])
p_chat_combine = chat_combine_template.replace("{summaries}", docs_together)
messages_combine = [{"role": "system", "content": p_chat_combine}]
source_log_docs = []
for doc in docs:
if doc.metadata:
data = json.dumps({"type": "source", "doc": doc.page_content, "metadata": doc.metadata})
source_log_docs.append({"title": doc.metadata['title'].split('/')[-1], "text": doc.page_content})
else:
data = json.dumps({"type": "source", "doc": doc.page_content})
source_log_docs.append({"title": doc.page_content, "text": doc.page_content})
yield f"data:{data}\n\n"
return 'Welcome to DocsGPT Backend!'
if len(chat_history) > 1:
tokens_current_history = 0
# count tokens in history
chat_history.reverse()
for i in chat_history:
if "prompt" in i and "response" in i:
tokens_batch = llm.get_num_tokens(i["prompt"]) + llm.get_num_tokens(i["response"])
if tokens_current_history + tokens_batch < settings.TOKENS_MAX_HISTORY:
tokens_current_history += tokens_batch
messages_combine.append({"role": "user", "content": i["prompt"]})
messages_combine.append({"role": "system", "content": i["response"]})
messages_combine.append({"role": "user", "content": question})
completion = openai.ChatCompletion.create(model=gpt_model, engine=settings.AZURE_DEPLOYMENT_NAME,
messages=messages_combine, stream=True, max_tokens=500, temperature=0)
reponse_full = ""
for line in completion:
if "content" in line["choices"][0]["delta"]:
# check if the delta contains content
data = json.dumps({"answer": str(line["choices"][0]["delta"]["content"])})
reponse_full += str(line["choices"][0]["delta"]["content"])
yield f"data: {data}\n\n"
# save conversation to database
if conversation_id is not None:
conversations_collection.update_one(
{"_id": ObjectId(conversation_id)},
{"$push": {"queries": {"prompt": question, "response": reponse_full, "sources": source_log_docs}}},
)
else:
# create new conversation
# generate summary
messages_summary = [{"role": "assistant", "content": "Summarise following conversation in no more than 3 "
"words, respond ONLY with the summary, use the same "
"language as the system \n\nUser: " + question + "\n\n" +
"AI: " +
reponse_full},
{"role": "user", "content": "Summarise following conversation in no more than 3 words, "
"respond ONLY with the summary, use the same language as the "
"system"}]
completion = openai.ChatCompletion.create(model='gpt-3.5-turbo', engine=settings.AZURE_DEPLOYMENT_NAME,
messages=messages_summary, max_tokens=30, temperature=0)
conversation_id = conversations_collection.insert_one(
{"user": "local",
"date": datetime.datetime.utcnow(),
"name": completion["choices"][0]["message"]["content"],
"queries": [{"prompt": question, "response": reponse_full, "sources": source_log_docs}]}
).inserted_id
# send data.type = "end" to indicate that the stream has ended as json
data = json.dumps({"type": "id", "id": str(conversation_id)})
yield f"data: {data}\n\n"
data = json.dumps({"type": "end"})
yield f"data: {data}\n\n"
@app.route("/stream", methods=["POST"])
def stream():
data = request.get_json()
# get parameter from url question
question = data["question"]
history = data["history"]
# history to json object from string
history = json.loads(history)
conversation_id = data["conversation_id"]
# check if active_docs is set
if not api_key_set:
api_key = data["api_key"]
else:
api_key = settings.API_KEY
if not embeddings_key_set:
embeddings_key = data["embeddings_key"]
else:
embeddings_key = settings.EMBEDDINGS_KEY
if "active_docs" in data:
vectorstore = get_vectorstore({"active_docs": data["active_docs"]})
else:
vectorstore = ""
docsearch = get_docsearch(vectorstore, embeddings_key)
# question = "Hi"
return Response(
complete_stream(question, docsearch,
chat_history=history, api_key=api_key,
conversation_id=conversation_id), mimetype="text/event-stream"
)
def is_azure_configured():
return settings.OPENAI_API_BASE and settings.OPENAI_API_VERSION and settings.AZURE_DEPLOYMENT_NAME
@app.route("/api/answer", methods=["POST"])
def api_answer():
data = request.get_json()
question = data["question"]
history = data["history"]
if "conversation_id" not in data:
conversation_id = None
else:
conversation_id = data["conversation_id"]
print("-" * 5)
if not api_key_set:
api_key = data["api_key"]
else:
api_key = settings.API_KEY
if not embeddings_key_set:
embeddings_key = data["embeddings_key"]
else:
embeddings_key = settings.EMBEDDINGS_KEY
# use try and except to check for exception
try:
# check if the vectorstore is set
vectorstore = get_vectorstore(data)
# loading the index and the store and the prompt template
# Note if you have used other embeddings than OpenAI, you need to change the embeddings
docsearch = get_docsearch(vectorstore, embeddings_key)
q_prompt = PromptTemplate(
input_variables=["context", "question"], template=template_quest, template_format="jinja2"
)
if settings.LLM_NAME == "openai_chat":
if is_azure_configured():
logger.debug("in Azure")
llm = AzureChatOpenAI(
openai_api_key=api_key,
openai_api_base=settings.OPENAI_API_BASE,
openai_api_version=settings.OPENAI_API_VERSION,
deployment_name=settings.AZURE_DEPLOYMENT_NAME,
)
else:
logger.debug("plain OpenAI")
llm = ChatOpenAI(openai_api_key=api_key, model_name=gpt_model) # optional parameter: model_name="gpt-4"
messages_combine = [SystemMessagePromptTemplate.from_template(chat_combine_template)]
if history:
tokens_current_history = 0
# count tokens in history
history.reverse()
for i in history:
if "prompt" in i and "response" in i:
tokens_batch = llm.get_num_tokens(i["prompt"]) + llm.get_num_tokens(i["response"])
if tokens_current_history + tokens_batch < settings.TOKENS_MAX_HISTORY:
tokens_current_history += tokens_batch
messages_combine.append(HumanMessagePromptTemplate.from_template(i["prompt"]))
messages_combine.append(AIMessagePromptTemplate.from_template(i["response"]))
messages_combine.append(HumanMessagePromptTemplate.from_template("{question}"))
p_chat_combine = ChatPromptTemplate.from_messages(messages_combine)
elif settings.LLM_NAME == "openai":
llm = OpenAI(openai_api_key=api_key, temperature=0)
elif settings.SELF_HOSTED_MODEL:
llm = hf
elif settings.LLM_NAME == "cohere":
llm = Cohere(model="command-xlarge-nightly", cohere_api_key=api_key)
else:
raise ValueError("unknown LLM model")
if settings.LLM_NAME == "openai_chat":
question_generator = LLMChain(llm=llm, prompt=CONDENSE_QUESTION_PROMPT)
doc_chain = load_qa_chain(llm, chain_type="map_reduce", combine_prompt=p_chat_combine)
chain = ConversationalRetrievalChain(
retriever=docsearch.as_retriever(k=2),
question_generator=question_generator,
combine_docs_chain=doc_chain,
)
chat_history = []
# result = chain({"question": question, "chat_history": chat_history})
# generate async with async generate method
result = run_async_chain(chain, question, chat_history)
elif settings.SELF_HOSTED_MODEL:
question_generator = LLMChain(llm=llm, prompt=CONDENSE_QUESTION_PROMPT)
doc_chain = load_qa_chain(llm, chain_type="map_reduce", combine_prompt=p_chat_combine)
chain = ConversationalRetrievalChain(
retriever=docsearch.as_retriever(k=2),
question_generator=question_generator,
combine_docs_chain=doc_chain,
)
chat_history = []
# result = chain({"question": question, "chat_history": chat_history})
# generate async with async generate method
result = run_async_chain(chain, question, chat_history)
else:
qa_chain = load_qa_chain(
llm=llm, chain_type="map_reduce", combine_prompt=chat_combine_template, question_prompt=q_prompt
)
chain = VectorDBQA(combine_documents_chain=qa_chain, vectorstore=docsearch, k=3)
result = chain({"query": question})
print(result)
# some formatting for the frontend
if "result" in result:
result["answer"] = result["result"]
result["answer"] = result["answer"].replace("\\n", "\n")
try:
result["answer"] = result["answer"].split("SOURCES:")[0]
except Exception:
pass
sources = docsearch.similarity_search(question, k=2)
sources_doc = []
for doc in sources:
if doc.metadata:
sources_doc.append({'title': doc.metadata['title'], 'text': doc.page_content})
else:
sources_doc.append({'title': doc.page_content, 'text': doc.page_content})
result['sources'] = sources_doc
# generate conversationId
if conversation_id is not None:
conversations_collection.update_one(
{"_id": ObjectId(conversation_id)},
{"$push": {"queries": {"prompt": question,
"response": result["answer"], "sources": result['sources']}}},
)
else:
# create new conversation
# generate summary
messages_summary = [AIMessage(content="Summarise following conversation in no more than 3 " +
"words, respond ONLY with the summary, use the same " +
"language as the system \n\nUser: " + question + "\n\nAI: " +
result["answer"]),
HumanMessage(content="Summarise following conversation in no more than 3 words, " +
"respond ONLY with the summary, use the same language as the " +
"system")]
# completion = openai.ChatCompletion.create(model='gpt-3.5-turbo', engine=settings.AZURE_DEPLOYMENT_NAME,
# messages=messages_summary, max_tokens=30, temperature=0)
completion = llm.predict_messages(messages_summary)
conversation_id = conversations_collection.insert_one(
{"user": "local",
"date": datetime.datetime.utcnow(),
"name": completion.content,
"queries": [{"prompt": question, "response": result["answer"], "sources": result['sources']}]}
).inserted_id
result["conversation_id"] = str(conversation_id)
# mock result
# result = {
# "answer": "The answer is 42",
# "sources": ["https://en.wikipedia.org/wiki/42_(number)", "https://en.wikipedia.org/wiki/42_(number)"]
# }
return result
except Exception as e:
# print whole traceback
traceback.print_exc()
print(str(e))
return bad_request(500, str(e))
@app.route("/api/docs_check", methods=["POST"])
def check_docs():
# check if docs exist in a vectorstore folder
data = request.get_json()
# split docs on / and take first part
if data["docs"].split("/")[0] == "local":
return {"status": "exists"}
vectorstore = "vectors/" + data["docs"]
base_path = "https://raw.githubusercontent.com/arc53/DocsHUB/main/"
if os.path.exists(vectorstore) or data["docs"] == "default":
return {"status": "exists"}
else:
r = requests.get(base_path + vectorstore + "index.faiss")
if r.status_code != 200:
return {"status": "null"}
else:
if not os.path.exists(vectorstore):
os.makedirs(vectorstore)
with open(vectorstore + "index.faiss", "wb") as f:
f.write(r.content)
# download the store
r = requests.get(base_path + vectorstore + "index.pkl")
with open(vectorstore + "index.pkl", "wb") as f:
f.write(r.content)
return {"status": "loaded"}
@app.route("/api/feedback", methods=["POST"])
def api_feedback():
data = request.get_json()
question = data["question"]
answer = data["answer"]
feedback = data["feedback"]
print("-" * 5)
print("Question: " + question)
print("Answer: " + answer)
print("Feedback: " + feedback)
print("-" * 5)
response = requests.post(
url="https://86x89umx77.execute-api.eu-west-2.amazonaws.com/docsgpt-feedback",
headers={
"Content-Type": "application/json; charset=utf-8",
},
data=json.dumps({"answer": answer, "question": question, "feedback": feedback}),
)
return {"status": http.client.responses.get(response.status_code, "ok")}
@app.route("/api/combine", methods=["GET"])
def combined_json():
user = "local"
"""Provide json file with combined available indexes."""
# get json from https://d3dg1063dc54p9.cloudfront.net/combined.json
data = [
{
"name": "default",
"language": "default",
"version": "",
"description": "default",
"fullName": "default",
"date": "default",
"docLink": "default",
"model": settings.EMBEDDINGS_NAME,
"location": "local",
}
]
# structure: name, language, version, description, fullName, date, docLink
# append data from vectors_collection
for index in vectors_collection.find({"user": user}):
data.append(
{
"name": index["name"],
"language": index["language"],
"version": "",
"description": index["name"],
"fullName": index["name"],
"date": index["date"],
"docLink": index["location"],
"model": settings.EMBEDDINGS_NAME,
"location": "local",
}
)
data_remote = requests.get("https://d3dg1063dc54p9.cloudfront.net/combined.json").json()
for index in data_remote:
index["location"] = "remote"
data.append(index)
return jsonify(data)
@app.route("/api/upload", methods=["POST"])
def upload_file():
"""Upload a file to get vectorized and indexed."""
if "user" not in request.form:
return {"status": "no user"}
user = secure_filename(request.form["user"])
if "name" not in request.form:
return {"status": "no name"}
job_name = secure_filename(request.form["name"])
# check if the post request has the file part
if "file" not in request.files:
print("No file part")
return {"status": "no file"}
file = request.files["file"]
if file.filename == "":
return {"status": "no file name"}
if file:
filename = secure_filename(file.filename)
# save dir
save_dir = os.path.join(app.config["UPLOAD_FOLDER"], user, job_name)
# create dir if not exists
if not os.path.exists(save_dir):
os.makedirs(save_dir)
file.save(os.path.join(save_dir, filename))
task = ingest.delay("temp", [".rst", ".md", ".pdf", ".txt"], job_name, filename, user)
# task id
task_id = task.id
return {"status": "ok", "task_id": task_id}
else:
return {"status": "error"}
@app.route("/api/task_status", methods=["GET"])
def task_status():
"""Get celery job status."""
task_id = request.args.get("task_id")
task = AsyncResult(task_id)
task_meta = task.info
return {"status": task.status, "result": task_meta}
### Backgound task api
@app.route("/api/upload_index", methods=["POST"])
def upload_index_files():
"""Upload two files(index.faiss, index.pkl) to the user's folder."""
if "user" not in request.form:
return {"status": "no user"}
user = secure_filename(request.form["user"])
if "name" not in request.form:
return {"status": "no name"}
job_name = secure_filename(request.form["name"])
if "file_faiss" not in request.files:
print("No file part")
return {"status": "no file"}
file_faiss = request.files["file_faiss"]
if file_faiss.filename == "":
return {"status": "no file name"}
if "file_pkl" not in request.files:
print("No file part")
return {"status": "no file"}
file_pkl = request.files["file_pkl"]
if file_pkl.filename == "":
return {"status": "no file name"}
# saves index files
save_dir = os.path.join("indexes", user, job_name)
if not os.path.exists(save_dir):
os.makedirs(save_dir)
file_faiss.save(os.path.join(save_dir, "index.faiss"))
file_pkl.save(os.path.join(save_dir, "index.pkl"))
# create entry in vectors_collection
vectors_collection.insert_one(
{
"user": user,
"name": job_name,
"language": job_name,
"location": save_dir,
"date": datetime.datetime.now().strftime("%d/%m/%Y %H:%M:%S"),
"model": settings.EMBEDDINGS_NAME,
"type": "local",
}
)
return {"status": "ok"}
@app.route("/api/download", methods=["get"])
def download_file():
user = secure_filename(request.args.get("user"))
job_name = secure_filename(request.args.get("name"))
filename = secure_filename(request.args.get("file"))
save_dir = os.path.join(app.config["UPLOAD_FOLDER"], user, job_name)
return send_from_directory(save_dir, filename, as_attachment=True)
@app.route("/api/delete_old", methods=["get"])
def delete_old():
"""Delete old indexes."""
import shutil
path = request.args.get("path")
dirs = path.split("/")
dirs_clean = []
for i in range(1, len(dirs)):
dirs_clean.append(secure_filename(dirs[i]))
# check that path strats with indexes or vectors
if dirs[0] not in ["indexes", "vectors"]:
return {"status": "error"}
path_clean = "/".join(dirs)
vectors_collection.delete_one({"location": path})
try:
shutil.rmtree(path_clean)
except FileNotFoundError:
pass
return {"status": "ok"}
@app.route("/api/get_conversations", methods=["get"])
def get_conversations():
# provides a list of conversations
conversations = conversations_collection.find().sort("date", -1)
list_conversations = []
for conversation in conversations:
list_conversations.append({"id": str(conversation["_id"]), "name": conversation["name"]})
#list_conversations = [{"id": "default", "name": "default"}, {"id": "jeff", "name": "jeff"}]
return jsonify(list_conversations)
@app.route("/api/get_single_conversation", methods=["get"])
def get_single_conversation():
# provides data for a conversation
conversation_id = request.args.get("id")
conversation = conversations_collection.find_one({"_id": ObjectId(conversation_id)})
return jsonify(conversation['queries'])
@app.route("/api/delete_conversation", methods=["POST"])
def delete_conversation():
# deletes a conversation from the database
conversation_id = request.args.get("id")
# write to mongodb
conversations_collection.delete_one(
{
"_id": ObjectId(conversation_id),
}
)
return {"status": "ok"}
# handling CORS
@app.after_request
def after_request(response):
response.headers.add("Access-Control-Allow-Origin", "*")
response.headers.add("Access-Control-Allow-Headers", "Content-Type,Authorization")
response.headers.add("Access-Control-Allow-Methods", "GET,PUT,POST,DELETE,OPTIONS")
response.headers.add("Access-Control-Allow-Credentials", "true")
return response
if __name__ == "__main__":
app.run(debug=True, port=7091)

9
application/celery.py Normal file
View File

@@ -0,0 +1,9 @@
from celery import Celery
from application.core.settings import settings
def make_celery(app_name=__name__):
celery = Celery(app_name, broker=settings.CELERY_BROKER_URL, backend=settings.CELERY_RESULT_BACKEND)
celery.conf.update(settings)
return celery
celery = make_celery()

View File

@@ -1,26 +1,43 @@
from pathlib import Path
from typing import Optional
import os
from pydantic import BaseSettings
from pydantic_settings import BaseSettings
current_dir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
class Settings(BaseSettings):
LLM_NAME: str = "openai_chat"
EMBEDDINGS_NAME: str = "openai_text-embedding-ada-002"
LLM_NAME: str = "docsgpt"
EMBEDDINGS_NAME: str = "huggingface_sentence-transformers/all-mpnet-base-v2"
CELERY_BROKER_URL: str = "redis://localhost:6379/0"
CELERY_RESULT_BACKEND: str = "redis://localhost:6379/1"
MONGO_URI: str = "mongodb://localhost:27017/docsgpt"
MODEL_PATH: str = "./models/gpt4all-model.bin"
MODEL_PATH: str = os.path.join(current_dir, "models/docsgpt-7b-f16.gguf")
TOKENS_MAX_HISTORY: int = 150
SELF_HOSTED_MODEL: bool = False
UPLOAD_FOLDER: str = "inputs"
VECTOR_STORE: str = "faiss" # "faiss" or "elasticsearch"
API_URL: str = "http://localhost:7091" # backend url for celery worker
API_KEY: str = None # LLM api key
EMBEDDINGS_KEY: str = None # api key for embeddings (if using openai, just copy API_KEY
OPENAI_API_BASE: str = None # azure openai api base url
OPENAI_API_VERSION: str = None # azure openai api version
AZURE_DEPLOYMENT_NAME: str = None # azure deployment name for answering
AZURE_EMBEDDINGS_DEPLOYMENT_NAME: str = None # azure deployment name for embeddings
API_KEY: Optional[str] = None # LLM api key
EMBEDDINGS_KEY: Optional[str] = None # api key for embeddings (if using openai, just copy API_KEY)
OPENAI_API_BASE: Optional[str] = None # azure openai api base url
OPENAI_API_VERSION: Optional[str] = None # azure openai api version
AZURE_DEPLOYMENT_NAME: Optional[str] = None # azure deployment name for answering
AZURE_EMBEDDINGS_DEPLOYMENT_NAME: Optional[str] = None # azure deployment name for embeddings
# elasticsearch
ELASTIC_CLOUD_ID: Optional[str] = None # cloud id for elasticsearch
ELASTIC_USERNAME: Optional[str] = None # username for elasticsearch
ELASTIC_PASSWORD: Optional[str] = None # password for elasticsearch
ELASTIC_URL: Optional[str] = None # url for elasticsearch
ELASTIC_INDEX: Optional[str] = "docsgpt" # index name for elasticsearch
# SageMaker config
SAGEMAKER_ENDPOINT: Optional[str] = None # SageMaker endpoint name
SAGEMAKER_REGION: Optional[str] = None # SageMaker region name
SAGEMAKER_ACCESS_KEY: Optional[str] = None # SageMaker access key
SAGEMAKER_SECRET_KEY: Optional[str] = None # SageMaker secret key
path = Path(__file__).parent.parent.absolute()

Binary file not shown.

Binary file not shown.

View File

View File

@@ -0,0 +1,40 @@
from application.llm.base import BaseLLM
from application.core.settings import settings
class AnthropicLLM(BaseLLM):
def __init__(self, api_key=None):
from anthropic import Anthropic, HUMAN_PROMPT, AI_PROMPT
self.api_key = api_key or settings.ANTHROPIC_API_KEY # If not provided, use a default from settings
self.anthropic = Anthropic(api_key=self.api_key)
self.HUMAN_PROMPT = HUMAN_PROMPT
self.AI_PROMPT = AI_PROMPT
def gen(self, model, messages, engine=None, max_tokens=300, stream=False, **kwargs):
context = messages[0]['content']
user_question = messages[-1]['content']
prompt = f"### Context \n {context} \n ### Question \n {user_question}"
if stream:
return self.gen_stream(model, prompt, max_tokens, **kwargs)
completion = self.anthropic.completions.create(
model=model,
max_tokens_to_sample=max_tokens,
stream=stream,
prompt=f"{self.HUMAN_PROMPT} {prompt}{self.AI_PROMPT}",
)
return completion.completion
def gen_stream(self, model, messages, engine=None, max_tokens=300, **kwargs):
context = messages[0]['content']
user_question = messages[-1]['content']
prompt = f"### Context \n {context} \n ### Question \n {user_question}"
stream_response = self.anthropic.completions.create(
model=model,
prompt=f"{self.HUMAN_PROMPT} {prompt}{self.AI_PROMPT}",
max_tokens_to_sample=max_tokens,
stream=True,
)
for completion in stream_response:
yield completion.completion

14
application/llm/base.py Normal file
View File

@@ -0,0 +1,14 @@
from abc import ABC, abstractmethod
class BaseLLM(ABC):
def __init__(self):
pass
@abstractmethod
def gen(self, *args, **kwargs):
pass
@abstractmethod
def gen_stream(self, *args, **kwargs):
pass

View File

@@ -0,0 +1,49 @@
from application.llm.base import BaseLLM
import json
import requests
class DocsGPTAPILLM(BaseLLM):
def __init__(self, *args, **kwargs):
self.endpoint = "https://llm.docsgpt.co.uk"
def gen(self, model, engine, messages, stream=False, **kwargs):
context = messages[0]['content']
user_question = messages[-1]['content']
prompt = f"### Instruction \n {user_question} \n ### Context \n {context} \n ### Answer \n"
response = requests.post(
f"{self.endpoint}/answer",
json={
"prompt": prompt,
"max_new_tokens": 30
}
)
response_clean = response.json()['a'].split("###")[0]
return response_clean
def gen_stream(self, model, engine, messages, stream=True, **kwargs):
context = messages[0]['content']
user_question = messages[-1]['content']
prompt = f"### Instruction \n {user_question} \n ### Context \n {context} \n ### Answer \n"
# send prompt to endpoint /stream
response = requests.post(
f"{self.endpoint}/stream",
json={
"prompt": prompt,
"max_new_tokens": 256
},
stream=True
)
for line in response.iter_lines():
if line:
#data = json.loads(line)
data_str = line.decode('utf-8')
if data_str.startswith("data: "):
data = json.loads(data_str[6:])
yield data['a']

View File

@@ -0,0 +1,44 @@
from application.llm.base import BaseLLM
class HuggingFaceLLM(BaseLLM):
def __init__(self, api_key, llm_name='Arc53/DocsGPT-7B',q=False):
global hf
from langchain.llms import HuggingFacePipeline
if q:
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline, BitsAndBytesConfig
tokenizer = AutoTokenizer.from_pretrained(llm_name)
bnb_config = BitsAndBytesConfig(
load_in_4bit=True,
bnb_4bit_use_double_quant=True,
bnb_4bit_quant_type="nf4",
bnb_4bit_compute_dtype=torch.bfloat16
)
model = AutoModelForCausalLM.from_pretrained(llm_name,quantization_config=bnb_config)
else:
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
tokenizer = AutoTokenizer.from_pretrained(llm_name)
model = AutoModelForCausalLM.from_pretrained(llm_name)
pipe = pipeline(
"text-generation", model=model,
tokenizer=tokenizer, max_new_tokens=2000,
device_map="auto", eos_token_id=tokenizer.eos_token_id
)
hf = HuggingFacePipeline(pipeline=pipe)
def gen(self, model, engine, messages, stream=False, **kwargs):
context = messages[0]['content']
user_question = messages[-1]['content']
prompt = f"### Instruction \n {user_question} \n ### Context \n {context} \n ### Answer \n"
result = hf(prompt)
return result.content
def gen_stream(self, model, engine, messages, stream=True, **kwargs):
raise NotImplementedError("HuggingFaceLLM Streaming is not implemented yet.")

View File

@@ -0,0 +1,39 @@
from application.llm.base import BaseLLM
from application.core.settings import settings
class LlamaCpp(BaseLLM):
def __init__(self, api_key, llm_name=settings.MODEL_PATH, **kwargs):
global llama
try:
from llama_cpp import Llama
except ImportError:
raise ImportError("Please install llama_cpp using pip install llama-cpp-python")
llama = Llama(model_path=llm_name, n_ctx=2048)
def gen(self, model, engine, messages, stream=False, **kwargs):
context = messages[0]['content']
user_question = messages[-1]['content']
prompt = f"### Instruction \n {user_question} \n ### Context \n {context} \n ### Answer \n"
result = llama(prompt, max_tokens=150, echo=False)
# import sys
# print(result['choices'][0]['text'].split('### Answer \n')[-1], file=sys.stderr)
return result['choices'][0]['text'].split('### Answer \n')[-1]
def gen_stream(self, model, engine, messages, stream=True, **kwargs):
context = messages[0]['content']
user_question = messages[-1]['content']
prompt = f"### Instruction \n {user_question} \n ### Context \n {context} \n ### Answer \n"
result = llama(prompt, max_tokens=150, echo=False, stream=stream)
# import sys
# print(list(result), file=sys.stderr)
for item in result:
for choice in item['choices']:
yield choice['text']

View File

@@ -0,0 +1,26 @@
from application.llm.openai import OpenAILLM, AzureOpenAILLM
from application.llm.sagemaker import SagemakerAPILLM
from application.llm.huggingface import HuggingFaceLLM
from application.llm.llama_cpp import LlamaCpp
from application.llm.anthropic import AnthropicLLM
from application.llm.docsgpt_provider import DocsGPTAPILLM
class LLMCreator:
llms = {
'openai': OpenAILLM,
'azure_openai': AzureOpenAILLM,
'sagemaker': SagemakerAPILLM,
'huggingface': HuggingFaceLLM,
'llama.cpp': LlamaCpp,
'anthropic': AnthropicLLM,
'docsgpt': DocsGPTAPILLM
}
@classmethod
def create_llm(cls, type, *args, **kwargs):
llm_class = cls.llms.get(type.lower())
if not llm_class:
raise ValueError(f"No LLM class found for type {type}")
return llm_class(*args, **kwargs)

60
application/llm/openai.py Normal file
View File

@@ -0,0 +1,60 @@
from application.llm.base import BaseLLM
from application.core.settings import settings
class OpenAILLM(BaseLLM):
def __init__(self, api_key):
global openai
from openai import OpenAI
self.client = OpenAI(
api_key=api_key,
)
self.api_key = api_key
def _get_openai(self):
# Import openai when needed
import openai
return openai
def gen(self, model, engine, messages, stream=False, **kwargs):
response = self.client.chat.completions.create(model=model,
messages=messages,
stream=stream,
**kwargs)
return response.choices[0].message.content
def gen_stream(self, model, engine, messages, stream=True, **kwargs):
response = self.client.chat.completions.create(model=model,
messages=messages,
stream=stream,
**kwargs)
for line in response:
# import sys
# print(line.choices[0].delta.content, file=sys.stderr)
if line.choices[0].delta.content is not None:
yield line.choices[0].delta.content
class AzureOpenAILLM(OpenAILLM):
def __init__(self, openai_api_key, openai_api_base, openai_api_version, deployment_name):
super().__init__(openai_api_key)
self.api_base = settings.OPENAI_API_BASE,
self.api_version = settings.OPENAI_API_VERSION,
self.deployment_name = settings.AZURE_DEPLOYMENT_NAME,
from openai import AzureOpenAI
self.client = AzureOpenAI(
api_key=openai_api_key,
api_version=settings.OPENAI_API_VERSION,
api_base=settings.OPENAI_API_BASE,
deployment_name=settings.AZURE_DEPLOYMENT_NAME,
)
def _get_openai(self):
openai = super()._get_openai()
return openai

View File

@@ -0,0 +1,139 @@
from application.llm.base import BaseLLM
from application.core.settings import settings
import json
import io
class LineIterator:
"""
A helper class for parsing the byte stream input.
The output of the model will be in the following format:
```
b'{"outputs": [" a"]}\n'
b'{"outputs": [" challenging"]}\n'
b'{"outputs": [" problem"]}\n'
...
```
While usually each PayloadPart event from the event stream will contain a byte array
with a full json, this is not guaranteed and some of the json objects may be split across
PayloadPart events. For example:
```
{'PayloadPart': {'Bytes': b'{"outputs": '}}
{'PayloadPart': {'Bytes': b'[" problem"]}\n'}}
```
This class accounts for this by concatenating bytes written via the 'write' function
and then exposing a method which will return lines (ending with a '\n' character) within
the buffer via the 'scan_lines' function. It maintains the position of the last read
position to ensure that previous bytes are not exposed again.
"""
def __init__(self, stream):
self.byte_iterator = iter(stream)
self.buffer = io.BytesIO()
self.read_pos = 0
def __iter__(self):
return self
def __next__(self):
while True:
self.buffer.seek(self.read_pos)
line = self.buffer.readline()
if line and line[-1] == ord('\n'):
self.read_pos += len(line)
return line[:-1]
try:
chunk = next(self.byte_iterator)
except StopIteration:
if self.read_pos < self.buffer.getbuffer().nbytes:
continue
raise
if 'PayloadPart' not in chunk:
print('Unknown event type:' + chunk)
continue
self.buffer.seek(0, io.SEEK_END)
self.buffer.write(chunk['PayloadPart']['Bytes'])
class SagemakerAPILLM(BaseLLM):
def __init__(self, *args, **kwargs):
import boto3
runtime = boto3.client(
'runtime.sagemaker',
aws_access_key_id='xxx',
aws_secret_access_key='xxx',
region_name='us-west-2'
)
self.endpoint = settings.SAGEMAKER_ENDPOINT
self.runtime = runtime
def gen(self, model, engine, messages, stream=False, **kwargs):
context = messages[0]['content']
user_question = messages[-1]['content']
prompt = f"### Instruction \n {user_question} \n ### Context \n {context} \n ### Answer \n"
# Construct payload for endpoint
payload = {
"inputs": prompt,
"stream": False,
"parameters": {
"do_sample": True,
"temperature": 0.1,
"max_new_tokens": 30,
"repetition_penalty": 1.03,
"stop": ["</s>", "###"]
}
}
body_bytes = json.dumps(payload).encode('utf-8')
# Invoke the endpoint
response = self.runtime.invoke_endpoint(EndpointName=self.endpoint,
ContentType='application/json',
Body=body_bytes)
result = json.loads(response['Body'].read().decode())
import sys
print(result[0]['generated_text'], file=sys.stderr)
return result[0]['generated_text'][len(prompt):]
def gen_stream(self, model, engine, messages, stream=True, **kwargs):
context = messages[0]['content']
user_question = messages[-1]['content']
prompt = f"### Instruction \n {user_question} \n ### Context \n {context} \n ### Answer \n"
# Construct payload for endpoint
payload = {
"inputs": prompt,
"stream": True,
"parameters": {
"do_sample": True,
"temperature": 0.1,
"max_new_tokens": 512,
"repetition_penalty": 1.03,
"stop": ["</s>", "###"]
}
}
body_bytes = json.dumps(payload).encode('utf-8')
# Invoke the endpoint
response = self.runtime.invoke_endpoint_with_response_stream(EndpointName=self.endpoint,
ContentType='application/json',
Body=body_bytes)
#result = json.loads(response['Body'].read().decode())
event_stream = response['Body']
start_json = b'{'
for line in LineIterator(event_stream):
if line != b'' and start_json in line:
#print(line)
data = json.loads(line[line.find(start_json):].decode('utf-8'))
if data['token']['text'] not in ["</s>", "###"]:
print(data['token']['text'],end='')
yield data['token']['text']

File diff suppressed because it is too large Load Diff

View File

@@ -1,5 +0,0 @@
{
"devDependencies": {
"tailwindcss": "^3.2.4"
}
}

View File

@@ -62,7 +62,6 @@ class SimpleDirectoryReader(BaseReader):
file_extractor: Optional[Dict[str, BaseParser]] = None,
num_files_limit: Optional[int] = None,
file_metadata: Optional[Callable[[str], Dict]] = None,
chunk_size_max: int = 2048,
) -> None:
"""Initialize with parameters."""
super().__init__()

View File

@@ -57,7 +57,7 @@ class HTMLParser(BaseParser):
title_indexes = [i for i, isd_el in enumerate(isd) if isd_el['type'] == 'Title']
# Creating 'Chunks' - List of lists of strings
# each list starting with with isd_el['type'] = 'Title' and all the data till the next 'Title'
# each list starting with isd_el['type'] = 'Title' and all the data till the next 'Title'
# Each Chunk can be thought of as an individual set of data, which can be sent to the model
# Where Each Title is grouped together with the data under it
@@ -69,10 +69,10 @@ class HTMLParser(BaseParser):
Chunks.append([])
Chunks[-1].append(isd_el['text'])
# Removing all the chunks with sum of lenth of all the strings in the chunk < 25
# Removing all the chunks with sum of length of all the strings in the chunk < 25
# TODO: This value can be an user defined variable
for chunk in Chunks:
# sum of lenth of all the strings in the chunk
# sum of length of all the strings in the chunk
sum = 0
sum += len(str(chunk))
if sum < 25:

View File

@@ -0,0 +1,51 @@
from urllib.parse import urlparse
from openapi_parser import parse
try:
from application.parser.file.base_parser import BaseParser
except ModuleNotFoundError:
from base_parser import BaseParser
class OpenAPI3Parser(BaseParser):
def init_parser(self) -> None:
return super().init_parser()
def get_base_urls(self, urls):
base_urls = []
for i in urls:
parsed_url = urlparse(i)
base_url = parsed_url.scheme + "://" + parsed_url.netloc
if base_url not in base_urls:
base_urls.append(base_url)
return base_urls
def get_info_from_paths(self, path):
info = ""
if path.operations:
for operation in path.operations:
info += (
f"\n{operation.method.value}="
f"{operation.responses[0].description}"
)
return info
def parse_file(self, file_path):
data = parse(file_path)
results = ""
base_urls = self.get_base_urls(link.url for link in data.servers)
base_urls = ",".join([base_url for base_url in base_urls])
results += f"Base URL:{base_urls}\n"
i = 1
for path in data.paths:
info = self.get_info_from_paths(path)
results += (
f"Path{i}: {path.url}\n"
f"description: {path.description}\n"
f"parameters: {path.parameters}\nmethods: {info}\n"
)
i += 1
with open("results.txt", "w") as f:
f.write(results)
return results

View File

@@ -27,7 +27,7 @@ class RstParser(BaseParser):
remove_interpreters: bool = True,
remove_directives: bool = True,
remove_whitespaces_excess: bool = True,
# Be carefull with remove_characters_excess, might cause data loss
# Be careful with remove_characters_excess, might cause data loss
remove_characters_excess: bool = True,
**kwargs: Any,
) -> None:

View File

@@ -1,14 +1,14 @@
import os
import tiktoken
from langchain.embeddings import OpenAIEmbeddings
from langchain.vectorstores import FAISS
from application.vectorstore.vector_creator import VectorCreator
from application.core.settings import settings
from retry import retry
# from langchain.embeddings import HuggingFaceEmbeddings
# from langchain.embeddings import HuggingFaceInstructEmbeddings
# from langchain.embeddings import CohereEmbeddings
# from langchain_community.embeddings import HuggingFaceEmbeddings
# from langchain_community.embeddings import HuggingFaceInstructEmbeddings
# from langchain_community.embeddings import CohereEmbeddings
def num_tokens_from_string(string: str, encoding_name: str) -> int:
@@ -33,12 +33,23 @@ def call_openai_api(docs, folder_name, task_status):
os.makedirs(f"{folder_name}")
from tqdm import tqdm
docs_test = [docs[0]]
docs.pop(0)
c1 = 0
if settings.VECTOR_STORE == "faiss":
docs_init = [docs[0]]
docs.pop(0)
store = FAISS.from_documents(docs_test, OpenAIEmbeddings(openai_api_key=os.getenv("EMBEDDINGS_KEY")))
store = VectorCreator.create_vectorstore(
settings.VECTOR_STORE,
docs_init = docs_init,
path=f"{folder_name}",
embeddings_key=os.getenv("EMBEDDINGS_KEY")
)
else:
store = VectorCreator.create_vectorstore(
settings.VECTOR_STORE,
path=f"{folder_name}",
embeddings_key=os.getenv("EMBEDDINGS_KEY")
)
# Uncomment for MPNet embeddings
# model_name = "sentence-transformers/all-mpnet-base-v2"
# hf = HuggingFaceEmbeddings(model_name=model_name)
@@ -57,7 +68,8 @@ def call_openai_api(docs, folder_name, task_status):
store.save_local(f"{folder_name}")
break
c1 += 1
store.save_local(f"{folder_name}")
if settings.VECTOR_STORE == "faiss":
store.save_local(f"{folder_name}")
def get_user_permission(docs, folder_name):

View File

@@ -0,0 +1,9 @@
You are a helpful AI assistant, DocsGPT, specializing in document assistance, designed to offer detailed and informative responses.
If appropriate, your answers can include code examples, formatted as follows:
```(language)
(code)
```
You effectively utilize chat history, ensuring relevant and tailored responses.
If a question doesn't align with your context, you provide friendly and helpful replies.
----------------
{summaries}

View File

@@ -0,0 +1,13 @@
You are an AI Assistant, DocsGPT, adept at offering document assistance.
Your expertise lies in providing answer on top of provided context.
You can leverage the chat history if needed.
Answer the question based on the context below.
Keep the answer concise. Respond "Irrelevant context" if not sure about the answer.
If question is not related to the context, respond "Irrelevant context".
When using code examples, use the following format:
```(language)
(code)
```
----------------
Context:
{summaries}

View File

@@ -1,25 +0,0 @@
You are a DocsGPT, friendly and helpful AI assistant by Arc53 that provides help with documents. You give thorough answers with code examples if possible.
QUESTION: How to merge tables in pandas?
=========
Content: pandas provides various facilities for easily combining together Series or DataFrame with various kinds of set logic for the indexes and relational algebra functionality in the case of join / merge-type operations.
Source: 28-pl
Content: pandas provides a single function, merge(), as the entry point for all standard database join operations between DataFrame or named Series objects: \n\npandas.merge(left, right, how='inner', on=None, left_on=None, right_on=None, left_index=False, right_index=False, sort=False, suffixes=('_x', '_y'), copy=True, indicator=False, validate=None)
Source: 30-pl
=========
FINAL ANSWER: To merge two tables in pandas, you can use the pd.merge() function. The basic syntax is: \n\npd.merge(left, right, on, how) \n\nwhere left and right are the two tables to merge, on is the column to merge on, and how is the type of merge to perform. \n\nFor example, to merge the two tables df1 and df2 on the column 'id', you can use: \n\npd.merge(df1, df2, on='id', how='inner')
SOURCES: 28-pl 30-pl
QUESTION: How are you?
=========
CONTENT:
SOURCE:
=========
FINAL ANSWER: I am fine, thank you. How are you?
SOURCES:
QUESTION: {{ question }}
=========
{{ summaries }}
=========
FINAL ANSWER:

View File

@@ -1,33 +0,0 @@
You are a DocsGPT, friendly and helpful AI assistant by Arc53 that provides help with documents. You give thorough answers with code examples if possible.
QUESTION: How to merge tables in pandas?
=========
Content: pandas provides various facilities for easily combining together Series or DataFrame with various kinds of set logic for the indexes and relational algebra functionality in the case of join / merge-type operations.
Source: 28-pl
Content: pandas provides a single function, merge(), as the entry point for all standard database join operations between DataFrame or named Series objects: \n\npandas.merge(left, right, how='inner', on=None, left_on=None, right_on=None, left_index=False, right_index=False, sort=False, suffixes=('_x', '_y'), copy=True, indicator=False, validate=None)
Source: 30-pl
=========
FINAL ANSWER: To merge two tables in pandas, you can use the pd.merge() function. The basic syntax is: \n\npd.merge(left, right, on, how) \n\nwhere left and right are the two tables to merge, on is the column to merge on, and how is the type of merge to perform. \n\nFor example, to merge the two tables df1 and df2 on the column 'id', you can use: \n\npd.merge(df1, df2, on='id', how='inner')
SOURCES: 28-pl 30-pl
QUESTION: How are you?
=========
CONTENT:
SOURCE:
=========
FINAL ANSWER: I am fine, thank you. How are you?
SOURCES:
QUESTION: {{ historyquestion }}
=========
CONTENT:
SOURCE:
=========
FINAL ANSWER: {{ historyanswer }}
SOURCES:
QUESTION: {{ question }}
=========
{{ summaries }}
=========
FINAL ANSWER:

View File

@@ -1,4 +0,0 @@
Use the following portion of a long document to see if any of the text is relevant to answer the question.
{{ context }}
Question: {{ question }}
Provide all relevant text to the question verbatim. Summarize if needed. If nothing relevant return "-".

View File

@@ -1,104 +1,33 @@
aiodns==3.0.0
aiohttp==3.8.5
aiohttp-retry==2.8.3
aiosignal==1.3.1
aleph-alpha-client==2.16.1
amqp==5.1.1
async-timeout==4.0.2
attrs==22.2.0
billiard==3.6.4.0
blobfile==2.0.1
boto3==1.28.20
celery==5.2.7
cffi==1.15.1
charset-normalizer==3.1.0
click==8.1.3
click-didyoumean==0.3.0
click-plugins==1.1.1
click-repl==0.2.0
cryptography==41.0.3
dataclasses-json==0.5.7
decorator==5.1.1
dill==0.3.6
dnspython==2.3.0
ecdsa==0.18.0
entrypoints==0.4
faiss-cpu==1.7.3
filelock==3.9.0
Flask==2.2.5
Flask-Cors==3.0.10
frozenlist==1.3.3
geojson==2.5.0
gunicorn==20.1.0
greenlet==2.0.2
gpt4all==0.1.7
huggingface-hub==0.15.1
humbug==0.3.2
idna==3.4
itsdangerous==2.1.2
Jinja2==3.1.2
jmespath==1.0.1
joblib==1.2.0
kombu==5.2.4
langchain==0.0.263
loguru==0.6.0
lxml==4.9.2
MarkupSafe==2.1.2
marshmallow==3.19.0
marshmallow-enum==1.5.1
mpmath==1.3.0
multidict==6.0.4
multiprocess==0.70.14
mypy-extensions==1.0.0
networkx==3.0
npx
anthropic==0.12.0
boto3==1.34.6
celery==5.3.6
dataclasses_json==0.6.3
docx2txt==0.8
EbookLib==0.18
elasticsearch==8.12.0
escodegen==1.0.11
esprima==4.0.1
faiss-cpu==1.7.4
Flask==3.0.1
gunicorn==21.2.0
html2text==2020.1.16
javalang==0.13.0
langchain==0.1.4
langchain-openai==0.0.5
nltk==3.8.1
numcodecs==0.11.0
numpy==1.24.2
openai==0.27.8
packaging==23.0
pathos==0.3.0
Pillow==9.4.0
pox==0.3.2
ppft==1.7.6.6
prompt-toolkit==3.0.38
py==1.11.0
pyasn1==0.4.8
pycares==4.3.0
pycparser==2.21
pycryptodomex==3.17
pydantic==1.10.5
PyJWT==2.6.0
pymongo==4.3.3
pyowm==3.3.0
openapi3_parser==1.1.16
pandas==2.2.0
pydantic_settings==2.1.0
pymongo==4.6.1
PyPDF2==3.0.1
PySocks==1.7.1
pytest
python-dateutil==2.8.2
python-dotenv==1.0.0
python-jose==3.3.0
pytz==2022.7.1
PyYAML==6.0
redis==4.5.4
regex==2022.10.31
requests==2.31.0
python-dotenv==1.0.1
redis==5.0.1
Requests==2.31.0
retry==0.9.2
rsa==4.9
scikit-learn==1.2.2
scipy==1.10.1
sentencepiece
six==1.16.0
SQLAlchemy==1.4.46
sympy==1.11.1
tenacity==8.2.2
threadpoolctl==3.1.0
tiktoken
tqdm==4.65.0
transformers==4.30.0
typer==0.7.0
typing-inspect==0.8.0
typing_extensions==4.5.0
urllib3==1.26.14
vine==5.0.0
wcwidth==0.2.6
yarl==1.8.2
sentence-transformers
tiktoken==0.5.2
torch==2.1.2
tqdm==4.66.1
transformers==4.36.2
unstructured==0.12.2
Werkzeug==3.0.1

Binary file not shown.

Before

Width:  |  Height:  |  Size: 37 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 352 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 34 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 631 B

Binary file not shown.

Before

Width:  |  Height:  |  Size: 1.7 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 15 KiB

View File

@@ -1 +0,0 @@
{"name":"","short_name":"","icons":[{"src":"/android-chrome-192x192.png","sizes":"192x192","type":"image/png"},{"src":"/android-chrome-512x512.png","sizes":"512x512","type":"image/png"}],"theme_color":"#ffffff","background_color":"#ffffff","display":"standalone"}

View File

@@ -1,19 +0,0 @@
function resetApiKey() {
const modal = document.getElementById("modal");
modal.classList.toggle("hidden");
}
const apiKeyForm = document.getElementById("api-key-form");
if (apiKeyForm) {
apiKeyForm.addEventListener("submit", function(event) {
event.preventDefault();
const apiKeyInput = document.getElementById("api-key-input");
const apiKey = apiKeyInput.value;
localStorage.setItem("apiKey", apiKey);
apiKeyInput.value = "";
modal.classList.toggle("hidden");
});
}

View File

@@ -1,76 +0,0 @@
var form = document.getElementById('message-form');
var errorModal = document.getElementById('error-alert')
document.getElementById('close').addEventListener('click',()=>{
errorModal.classList.toggle('hidden')
})
function submitForm(event){
event.preventDefault()
var message = document.getElementById("message-input").value;
console.log(message.length)
if(message.length === 0){
return
}
msg_html = '<div class="bg-blue-500 text-white p-2 rounded-lg mb-2 self-end"><p class="text-sm">'
msg_html += message
msg_html += '</p></div>'
document.getElementById("messages").innerHTML += msg_html;
let chatWindow = document.getElementById("messages-container");
chatWindow.scrollTop = chatWindow.scrollHeight;
document.getElementById("message-input").value = "";
document.getElementById("button-submit").innerHTML = '<i class="fa fa-circle-o-notch fa-spin"></i> Thinking...';
document.getElementById("button-submit").disabled = true;
if (localStorage.getItem('activeDocs') == null) {
localStorage.setItem('activeDocs', 'default')
}
fetch('/api/answer', {
method: 'POST',
headers: {
'Content-Type': 'application/json',
},
body: JSON.stringify({question: message,
api_key: localStorage.getItem('apiKey'),
embeddings_key: localStorage.getItem('apiKey'),
history: localStorage.getItem('chatHistory'),
active_docs: localStorage.getItem('activeDocs')}),
}).then((response)=> response.json())
.then(data => {
console.log('Success:', data);
if(data.error){
document.getElementById('text-error').textContent = `Error : ${JSON.stringify(data.message)}`
errorModal.classList.toggle('hidden')
}
if(data.answer){
msg_html = '<div class="bg-indigo-500 text-white p-2 rounded-lg mb-2 self-start"><code class="text-sm">'
data.answer = data.answer.replace(/\n/g, "<br>");
msg_html += data.answer
msg_html += '</code></div>'
document.getElementById("messages").innerHTML += msg_html;
let chatWindow = document.getElementById("messages-container");
chatWindow.scrollTop = chatWindow.scrollHeight;
}
document.getElementById("button-submit").innerHTML = 'Send';
document.getElementById("button-submit").disabled = false;
let chatHistory = [message, data.answer || ''];
localStorage.setItem('chatHistory', JSON.stringify(chatHistory));
})
.catch((error) => {
console.error('Error:', error);
// console.log(error);
// document.getElementById("button-submit").innerHTML = 'Send';
// document.getElementById("button-submit").disabled = false;
});
}
//window.addEventListener('submit',submitForm)
// rewrite using id = button-submit
document.getElementById("button-submit").addEventListener('click',submitForm)

View File

@@ -1,15 +0,0 @@
document.getElementById("select-docs").addEventListener("change", function() {
localStorage.setItem('activeDocs', this.value)
fetch('/api/docs_check', {
method: 'POST',
headers: {
'Content-Type': 'application/json',
},
body: JSON.stringify({docs: this.value}),
}).then(response => response.json()).then(
data => {
console.log('Success:', data);
}
)
});

View File

@@ -1,37 +0,0 @@
@tailwind base;
@tailwind components;
@tailwind utilities;
@media screen and (max-width: 1024px) {
.text-lg {
font-size: 3.125rem;
margin: 2rem;
line-height: inherit;
}
.text-sm {
font-size: 2.5rem;
margin: 1.5rem;
line-height: inherit;
}
}
.loader {
border: 16px solid #f3f3f3; /* Light grey */
border-top: 16px solid #3498db; /* Blue */
border-radius: 50%;
width: 120px;
height: 120px;
animation: spin 2s linear infinite;
}
@keyframes spin {
0% { transform: rotate(0deg); }
100% { transform: rotate(360deg); }
}

View File

@@ -1,215 +0,0 @@
<!DOCTYPE html>
<html>
<head>
<title>DocsGPT 🦖 Preview</title>
<link href="{{url_for('static',filename='dist/css/output.css')}}" rel="stylesheet">
<link rel="favicon" href="{{ url_for('static', filename='favicon/favicon.ico') }}">
<link rel="apple-touch-icon" sizes="180x180" href="{{ url_for('static', filename='favicon/apple-touch-icon.png') }}">
<link rel="icon" type="image/png" sizes="32x32" href="{{ url_for('static', filename='favicon/favicon-32x32.png') }}">
<link rel="icon" type="image/png" sizes="16x16" href="{{ url_for('static', filename='favicon/favicon-16x16.png') }}">
<link rel="manifest" href="{{ url_for('static', filename='favicon//site.webmanifest') }}">
<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/font-awesome/4.7.0/css/font-awesome.min.css">
</head>
<body>
<header class="bg-white p-2 flex justify-between items-center">
<h1 class="text-lg font-medium">DocsGPT 🦖 Preview</h1>
<div>
<a href="https://github.com/arc53/docsgpt" class="text-blue-500 hover:text-blue-800 text-sm">About</a>
{% if not api_key_set %}
<button class="text-sm text-yellow-500 hover:text-yellow-800" onclick="resetApiKey()">Reset Key</button>
{% endif %}
</div>
</header>
<!-- Alert Info -->
<div class="border flex justify-between
w-auto px-4 py-3 rounded relative
hidden" style="background-color: rgb(197, 51, 51);color: white;" id="error-alert" role="alert">
<span class="block sm:inline" id="text-error"></span>
<strong class="text-xl align-center alert-del" style="cursor: pointer;" id="close">&times;</strong>
</div>
<div class="lg:flex ml-2 mr-2">
<div class="lg:w-3/4 min-h-screen max-h-screen">
<div class="w-full flex flex-col h-5/6">
<div id="messages-container" style="overflow: auto;" class="sm:max-lg:mb-[12rem]">
<div id="messages" class="w-full flex flex-col mt-2" >
<div class="bg-indigo-500 text-white p-2 rounded-lg mb-2 self-start">
<p class="text-sm">Hello, ask me anything about this library. Im here to help</p>
</div>
<div class="bg-blue-500 text-white p-2 rounded-lg mb-2 self-end">
<p class="text-sm">How to merge tables?</p>
</div>
<div class="bg-indigo-500 text-white p-2 rounded-lg mb-2 self-start">
<p class="text-sm">To merge two tables in pandas, you can use the pd.merge() function. The basic syntax is:<br>
pd.merge(left, right, on, how)<br>
where left and right are the two tables to merge, on is the column to merge on, and how is the type of merge to perform.<br>
For example, to merge the two tables df1 and df2 on the column 'key', you can use:<br>
pd.merge(df1, df2, on='key', how='left')<br>
This will return a new DataFrame with all the columns from both tables, and only the rows that match the 'key' column. </p>
</div>
</div>
</div>
<div class="fixed bottom-0 w-full mt-4 mb-2 lg:w-3/4">
<form id="message-form" autocomplete="off" class="flex items-stretch">
<input autocomplete="off" id="message-input" class="bg-white p-2 rounded-lg ml-2 text-sm w-full" type="text" placeholder="Type your message here...">
<button id="button-submit" class="bg-blue-500 text-white p-2 rounded-lg ml-2 mr-2 text-sm sm:max-lg:p-5" type="submit">Send</button>
</form>
</div>
</div>
</div>
<div class="lg:w-1/4 p-2 sm:max-lg:hidden">
<p class="text-sm">This is a chatbot that uses the GPT-3, Faiss and <a href="https://github.com/hwchase17/langchain" class="text-blue-500 hover:text-blue-800">LangChain</a> to answer questions</p>
<br>
<p class="text-sm">The source code is available on <a href="https://github.com/arc53/docsgpt" class="text-blue-500 hover:text-blue-800">Github</a></p><br>
<p class="text-sm">Currently It uses python pandas documentation, so it will respond to information relevant to pandas. If you want to train it on different documentation - <a href="https://github.com/arc53/docsgpt/wiki/How-to-train-on-other-documentation" class="text-blue-500 hover:text-blue-800"> please follow this guide </a></p><br>
<p class="text-sm">If you want to launch it on your own server - <a href="https://github.com/arc53/docsgpt/wiki/How-to-train-on-other-documentation" class="text-blue-500 hover:text-blue-800"> follow this guide </a></p><br>
<label class="block mb-2 text-sm font-medium text-gray-900">Select documentation from DocsHUB</label>
<select id="select-docs" class="bg-gray-50 border border-gray-300 text-gray-900 text-sm rounded-lg focus:ring-blue-500 focus:border-blue-500 block w-full p-2.5">
<option selected value="default">Choose documentation</option>
<option value="default">Default</option>
</select>
<form action="/api/upload" method="post" enctype="multipart/form-data" class="mt-2">
<input type="file" name="file" class="py-4" id="file-upload">
<input type="text" name="user" value="local" hidden>
<input type="text" name="name" placeholder="Name:">
<button type="submit" class="py-2 px-4 text-white bg-blue-500 rounded-md hover:bg-blue-600 focus:outline-none focus:ring-2 focus:ring-offset-2 focus:ring-blue-500">
Upload
</button>
</form>
</div>
</div>
<div class="flex items-center justify-center h-full">
</div>
{% if not api_key_set %}
<div class="fixed z-10 overflow-y-auto top-0 w-full left-0 show" id="modal">
<div class="flex items-center justify-center min-height-100vh pt-4 px-4 pb-20 text-center sm:block sm:p-0">
<div class="fixed inset-0 transition-opacity">
<div class="absolute inset-0 bg-gray-900 opacity-75" />
</div>
<span class="hidden sm:inline-block sm:align-middle sm:h-screen">&#8203;</span>
<div class=" text-sm inline-block align-center bg-white rounded-lg text-left overflow-hidden shadow-xl transform transition-all sm:my-8 sm:align-middle sm:max-w-lg sm:w-full" role="dialog" aria-modal="true" aria-labelledby="modal-headline">
<form id="api-key-form">
<div class="bg-white px-4 pt-5 pb-4 sm:p-6 sm:pb-4">
<h2>Before you can start using DocsGPT we need you to provide an API key for llm. Currently, we support only OpenAI but soon many more. You can find it <a class="text-blue-500 hover:text-blue-800" href="https://platform.openai.com/account/api-keys">here</a></h2><br>
<label>OpenAI API key:</label>
<input id="api-key-input" type="password" class="w-full bg-gray-100 p-2 mt-2 mb-3" placeholder="Paste you Api Key here">
</div>
<div class="bg-gray-200 px-4 py-3 text-right">
<button type="submit" class="py-2 px-4 bg-blue-500 text-white rounded hover:bg-blue-700 mr-2">Save</button>
</div>
</form>
</div>
</div>
</div>
{% endif %}
<script>
function docsIndex() {
// loads latest index from https://raw.githubusercontent.com/arc53/DocsHUB/main/combined.json
// and stores it in localStorage
fetch('/api/combine')
.then(response => response.json())
.then(data => {
localStorage.setItem("docsIndex", JSON.stringify(data));
localStorage.setItem("docsIndexDate", Date.now());
generateOptions()
}
)
}
function generateOptions(){
docsIndex = localStorage.getItem('docsIndex')
// create option on select with id select-docs
var select = document.getElementById("select-docs");
// convert docsIndex to json
docsIndex = JSON.parse(docsIndex)
// create option for each key in docsIndex
for (var key in docsIndex) {
var option = document.createElement("option");
if (docsIndex[key].location == 'docshub'){
if (docsIndex[key].name == docsIndex[key].language) {
option.text = docsIndex[key].name + " " + docsIndex[key].version;
option.value = docsIndex[key].name + "/" + ".project" + "/" + docsIndex[key].version + "/{{ embeddings_choice }}/";
if (docsIndex[key].model == "{{ embeddings_choice }}") {
select.add(option);
}
}
else {
option.text = docsIndex[key].name + " " + docsIndex[key].version;
option.value = docsIndex[key].language + "/" + docsIndex[key].name + "/" + docsIndex[key].version + "/{{ embeddings_choice }}/";
if (docsIndex[key].model == "{{ embeddings_choice }}") {
select.add(option);
}
}
}
else {
option.text = docsIndex[key].name;
option.value = docsIndex[key].location + "/" + docsIndex[key].name;
select.add(option);
}
}
}
{% if not api_key_set %}
if (localStorage.getItem('apiKey') === null) {
console.log("apiKey is not set")
document.getElementById('modal').classList.toggle('hidden')
}
{% endif %}
if (localStorage.getItem('docsIndex') === null) {
console.log("docsIndex is not set")
docsIndex()
}
else if (localStorage.getItem("docsIndexDate") < Date.now() - 900000) {
console.log("docsIndex is older than 15 minutes")
docsIndex()
}
generateOptions()
</script>
{% if not api_key_set %}
<script src="{{url_for('static',filename='src/authapi.js')}}"></script>
{% endif %}
<script src="{{url_for('static',filename='src/chat.js')}}"></script>
<script src="{{url_for('static',filename='src/choiceChange.js')}}"></script>
</body>
</html>

View File

View File

@@ -0,0 +1,56 @@
from abc import ABC, abstractmethod
import os
from langchain_community.embeddings import (
HuggingFaceEmbeddings,
CohereEmbeddings,
HuggingFaceInstructEmbeddings,
)
from langchain_openai import OpenAIEmbeddings
from application.core.settings import settings
class BaseVectorStore(ABC):
def __init__(self):
pass
@abstractmethod
def search(self, *args, **kwargs):
pass
def is_azure_configured(self):
return settings.OPENAI_API_BASE and settings.OPENAI_API_VERSION and settings.AZURE_DEPLOYMENT_NAME
def _get_embeddings(self, embeddings_name, embeddings_key=None):
embeddings_factory = {
"openai_text-embedding-ada-002": OpenAIEmbeddings,
"huggingface_sentence-transformers/all-mpnet-base-v2": HuggingFaceEmbeddings,
"huggingface_hkunlp/instructor-large": HuggingFaceInstructEmbeddings,
"cohere_medium": CohereEmbeddings
}
if embeddings_name not in embeddings_factory:
raise ValueError(f"Invalid embeddings_name: {embeddings_name}")
if embeddings_name == "openai_text-embedding-ada-002":
if self.is_azure_configured():
os.environ["OPENAI_API_TYPE"] = "azure"
embedding_instance = embeddings_factory[embeddings_name](
model=settings.AZURE_EMBEDDINGS_DEPLOYMENT_NAME
)
else:
embedding_instance = embeddings_factory[embeddings_name](
openai_api_key=embeddings_key
)
elif embeddings_name == "cohere_medium":
embedding_instance = embeddings_factory[embeddings_name](
cohere_api_key=embeddings_key
)
elif embeddings_name == "huggingface_sentence-transformers/all-mpnet-base-v2":
embedding_instance = embeddings_factory[embeddings_name](
#model_name="./model/all-mpnet-base-v2",
model_kwargs={"device": "cpu"},
)
else:
embedding_instance = embeddings_factory[embeddings_name]()
return embedding_instance

View File

@@ -0,0 +1,8 @@
class Document(str):
"""Class for storing a piece of text and associated metadata."""
def __new__(cls, page_content: str, metadata: dict):
instance = super().__new__(cls, page_content)
instance.page_content = page_content
instance.metadata = metadata
return instance

View File

@@ -0,0 +1,213 @@
from application.vectorstore.base import BaseVectorStore
from application.core.settings import settings
from application.vectorstore.document_class import Document
import elasticsearch
class ElasticsearchStore(BaseVectorStore):
_es_connection = None # Class attribute to hold the Elasticsearch connection
def __init__(self, path, embeddings_key, index_name=settings.ELASTIC_INDEX):
super().__init__()
self.path = path.replace("application/indexes/", "").rstrip("/")
self.embeddings_key = embeddings_key
self.index_name = index_name
if ElasticsearchStore._es_connection is None:
connection_params = {}
if settings.ELASTIC_URL:
connection_params["hosts"] = [settings.ELASTIC_URL]
connection_params["http_auth"] = (settings.ELASTIC_USERNAME, settings.ELASTIC_PASSWORD)
elif settings.ELASTIC_CLOUD_ID:
connection_params["cloud_id"] = settings.ELASTIC_CLOUD_ID
connection_params["basic_auth"] = (settings.ELASTIC_USERNAME, settings.ELASTIC_PASSWORD)
else:
raise ValueError("Please provide either elasticsearch_url or cloud_id.")
ElasticsearchStore._es_connection = elasticsearch.Elasticsearch(**connection_params)
self.docsearch = ElasticsearchStore._es_connection
def connect_to_elasticsearch(
*,
es_url = None,
cloud_id = None,
api_key = None,
username = None,
password = None,
):
try:
import elasticsearch
except ImportError:
raise ImportError(
"Could not import elasticsearch python package. "
"Please install it with `pip install elasticsearch`."
)
if es_url and cloud_id:
raise ValueError(
"Both es_url and cloud_id are defined. Please provide only one."
)
connection_params = {}
if es_url:
connection_params["hosts"] = [es_url]
elif cloud_id:
connection_params["cloud_id"] = cloud_id
else:
raise ValueError("Please provide either elasticsearch_url or cloud_id.")
if api_key:
connection_params["api_key"] = api_key
elif username and password:
connection_params["basic_auth"] = (username, password)
es_client = elasticsearch.Elasticsearch(
**connection_params,
)
try:
es_client.info()
except Exception as e:
raise e
return es_client
def search(self, question, k=2, index_name=settings.ELASTIC_INDEX, *args, **kwargs):
embeddings = self._get_embeddings(settings.EMBEDDINGS_NAME, self.embeddings_key)
vector = embeddings.embed_query(question)
knn = {
"filter": [{"match": {"metadata.store.keyword": self.path}}],
"field": "vector",
"k": k,
"num_candidates": 100,
"query_vector": vector,
}
full_query = {
"knn": knn,
"query": {
"bool": {
"must": [
{
"match": {
"text": {
"query": question,
}
}
}
],
"filter": [{"match": {"metadata.store.keyword": self.path}}],
}
},
"rank": {"rrf": {}},
}
resp = self.docsearch.search(index=self.index_name, query=full_query['query'], size=k, knn=full_query['knn'])
# create Documents objects from the results page_content ['_source']['text'], metadata ['_source']['metadata']
doc_list = []
for hit in resp['hits']['hits']:
doc_list.append(Document(page_content = hit['_source']['text'], metadata = hit['_source']['metadata']))
return doc_list
def _create_index_if_not_exists(
self, index_name, dims_length
):
if self._es_connection.indices.exists(index=index_name):
print(f"Index {index_name} already exists.")
else:
indexSettings = self.index(
dims_length=dims_length,
)
self._es_connection.indices.create(index=index_name, **indexSettings)
def index(
self,
dims_length,
):
return {
"mappings": {
"properties": {
"vector": {
"type": "dense_vector",
"dims": dims_length,
"index": True,
"similarity": "cosine",
},
}
}
}
def add_texts(
self,
texts,
metadatas = None,
ids = None,
refresh_indices = True,
create_index_if_not_exists = True,
bulk_kwargs = None,
**kwargs,
):
from elasticsearch.helpers import BulkIndexError, bulk
bulk_kwargs = bulk_kwargs or {}
import uuid
embeddings = []
ids = ids or [str(uuid.uuid4()) for _ in texts]
requests = []
embeddings = self._get_embeddings(settings.EMBEDDINGS_NAME, self.embeddings_key)
vectors = embeddings.embed_documents(list(texts))
dims_length = len(vectors[0])
if create_index_if_not_exists:
self._create_index_if_not_exists(
index_name=self.index_name, dims_length=dims_length
)
for i, (text, vector) in enumerate(zip(texts, vectors)):
metadata = metadatas[i] if metadatas else {}
requests.append(
{
"_op_type": "index",
"_index": self.index_name,
"text": text,
"vector": vector,
"metadata": metadata,
"_id": ids[i],
}
)
if len(requests) > 0:
try:
success, failed = bulk(
self._es_connection,
requests,
stats_only=True,
refresh=refresh_indices,
**bulk_kwargs,
)
return ids
except BulkIndexError as e:
print(f"Error adding texts: {e}")
firstError = e.errors[0].get("index", {}).get("error", {})
print(f"First error reason: {firstError.get('reason')}")
raise e
else:
return []
def delete_index(self):
self._es_connection.delete_by_query(index=self.index_name, query={"match": {
"metadata.store.keyword": self.path}},)

View File

@@ -0,0 +1,46 @@
from langchain_community.vectorstores import FAISS
from application.vectorstore.base import BaseVectorStore
from application.core.settings import settings
class FaissStore(BaseVectorStore):
def __init__(self, path, embeddings_key, docs_init=None):
super().__init__()
self.path = path
embeddings = self._get_embeddings(settings.EMBEDDINGS_NAME, embeddings_key)
if docs_init:
self.docsearch = FAISS.from_documents(
docs_init, embeddings
)
else:
self.docsearch = FAISS.load_local(
self.path, embeddings
)
self.assert_embedding_dimensions(embeddings)
def search(self, *args, **kwargs):
return self.docsearch.similarity_search(*args, **kwargs)
def add_texts(self, *args, **kwargs):
return self.docsearch.add_texts(*args, **kwargs)
def save_local(self, *args, **kwargs):
return self.docsearch.save_local(*args, **kwargs)
def delete_index(self, *args, **kwargs):
return self.docsearch.delete(*args, **kwargs)
def assert_embedding_dimensions(self, embeddings):
"""
Check that the word embedding dimension of the docsearch index matches
the dimension of the word embeddings used
"""
if settings.EMBEDDINGS_NAME == "huggingface_sentence-transformers/all-mpnet-base-v2":
try:
word_embedding_dimension = embeddings.client[1].word_embedding_dimension
except AttributeError as e:
raise AttributeError("word_embedding_dimension not found in embeddings.client[1]") from e
docsearch_index_dimension = self.docsearch.index.d
if word_embedding_dimension != docsearch_index_dimension:
raise ValueError(f"word_embedding_dimension ({word_embedding_dimension}) " +
f"!= docsearch_index_word_embedding_dimension ({docsearch_index_dimension})")

View File

@@ -0,0 +1,126 @@
from application.vectorstore.base import BaseVectorStore
from application.core.settings import settings
from application.vectorstore.document_class import Document
class MongoDBVectorStore(BaseVectorStore):
def __init__(
self,
path: str = "",
embeddings_key: str = "embeddings",
collection: str = "documents",
index_name: str = "vector_search_index",
text_key: str = "text",
embedding_key: str = "embedding",
database: str = "docsgpt",
):
self._index_name = index_name
self._text_key = text_key
self._embedding_key = embedding_key
self._embeddings_key = embeddings_key
self._mongo_uri = settings.MONGO_URI
self._path = path.replace("application/indexes/", "").rstrip("/")
self._embedding = self._get_embeddings(settings.EMBEDDINGS_NAME, embeddings_key)
try:
import pymongo
except ImportError:
raise ImportError(
"Could not import pymongo python package. "
"Please install it with `pip install pymongo`."
)
self._client = pymongo.MongoClient(self._mongo_uri)
self._database = self._client[database]
self._collection = self._database[collection]
def search(self, question, k=2, *args, **kwargs):
query_vector = self._embedding.embed_query(question)
pipeline = [
{
"$vectorSearch": {
"queryVector": query_vector,
"path": self._embedding_key,
"limit": k,
"numCandidates": k * 10,
"index": self._index_name,
"filter": {
"store": {"$eq": self._path}
}
}
}
]
cursor = self._collection.aggregate(pipeline)
results = []
for doc in cursor:
text = doc[self._text_key]
doc.pop("_id")
doc.pop(self._text_key)
doc.pop(self._embedding_key)
metadata = doc
results.append(Document(text, metadata))
return results
def _insert_texts(self, texts, metadatas):
if not texts:
return []
embeddings = self._embedding.embed_documents(texts)
to_insert = [
{self._text_key: t, self._embedding_key: embedding, **m}
for t, m, embedding in zip(texts, metadatas, embeddings)
]
# insert the documents in MongoDB Atlas
insert_result = self._collection.insert_many(to_insert)
return insert_result.inserted_ids
def add_texts(self,
texts,
metadatas = None,
ids = None,
refresh_indices = True,
create_index_if_not_exists = True,
bulk_kwargs = None,
**kwargs,):
#dims = self._embedding.client[1].word_embedding_dimension
# # check if index exists
# if create_index_if_not_exists:
# # check if index exists
# info = self._collection.index_information()
# if self._index_name not in info:
# index_mongo = {
# "fields": [{
# "type": "vector",
# "path": self._embedding_key,
# "numDimensions": dims,
# "similarity": "cosine",
# },
# {
# "type": "filter",
# "path": "store"
# }]
# }
# self._collection.create_index(self._index_name, index_mongo)
batch_size = 100
_metadatas = metadatas or ({} for _ in texts)
texts_batch = []
metadatas_batch = []
result_ids = []
for i, (text, metadata) in enumerate(zip(texts, _metadatas)):
texts_batch.append(text)
metadatas_batch.append(metadata)
if (i + 1) % batch_size == 0:
result_ids.extend(self._insert_texts(texts_batch, metadatas_batch))
texts_batch = []
metadatas_batch = []
if texts_batch:
result_ids.extend(self._insert_texts(texts_batch, metadatas_batch))
return result_ids
def delete_index(self, *args, **kwargs):
self._collection.delete_many({"store": self._path})

View File

@@ -0,0 +1,18 @@
from application.vectorstore.faiss import FaissStore
from application.vectorstore.elasticsearch import ElasticsearchStore
from application.vectorstore.mongodb import MongoDBVectorStore
class VectorCreator:
vectorstores = {
'faiss': FaissStore,
'elasticsearch':ElasticsearchStore,
'mongodb': MongoDBVectorStore,
}
@classmethod
def create_vectorstore(cls, type, *args, **kwargs):
vectorstore_class = cls.vectorstores.get(type.lower())
if not vectorstore_class:
raise ValueError(f"No vectorstore class found for type {type}")
return vectorstore_class(*args, **kwargs)

View File

@@ -20,15 +20,34 @@ except FileExistsError:
pass
# Define a function to extract metadata from a given filename.
def metadata_from_filename(title):
return {'title': title}
store = '/'.join(title.split('/')[1:3])
return {'title': title, 'store': store}
# Define a function to generate a random string of a given length.
def generate_random_string(length):
return ''.join([string.ascii_letters[i % 52] for i in range(length)])
current_dir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
# Define the main function for ingesting and processing documents.
def ingest_worker(self, directory, formats, name_job, filename, user):
"""
Ingest and process documents.
Args:
self: Reference to the instance of the task.
directory (str): Specifies the directory for ingesting ('inputs' or 'temp').
formats (list of str): List of file extensions to consider for ingestion (e.g., [".rst", ".md"]).
name_job (str): Name of the job for this ingestion task.
filename (str): Name of the file to be ingested.
user (str): Identifier for the user initiating the ingestion.
Returns:
dict: Information about the completed ingestion task, including input parameters and a "limited" flag.
"""
# directory = 'inputs' or 'temp'
# formats = [".rst", ".md"]
input_files = None
@@ -43,9 +62,13 @@ def ingest_worker(self, directory, formats, name_job, filename, user):
min_tokens = 150
max_tokens = 1250
full_path = directory + '/' + user + '/' + name_job
import sys
print(full_path, file=sys.stderr)
# check if API_URL env variable is set
file_data = {'name': name_job, 'file': filename, 'user': user}
response = requests.get(urljoin(settings.API_URL, "/api/download"), params=file_data)
# check if file is in the response
print(response, file=sys.stderr)
file = response.content
if not os.path.exists(full_path):
@@ -78,11 +101,15 @@ def ingest_worker(self, directory, formats, name_job, filename, user):
# get files from outputs/inputs/index.faiss and outputs/inputs/index.pkl
# and send them to the server (provide user and name in form)
file_data = {'name': name_job, 'user': user}
files = {'file_faiss': open(full_path + '/index.faiss', 'rb'),
'file_pkl': open(full_path + '/index.pkl', 'rb')}
response = requests.post(urljoin(settings.API_URL, "/api/upload_index"), files=files, data=file_data)
if settings.VECTOR_STORE == "faiss":
files = {'file_faiss': open(full_path + '/index.faiss', 'rb'),
'file_pkl': open(full_path + '/index.pkl', 'rb')}
response = requests.post(urljoin(settings.API_URL, "/api/upload_index"), files=files, data=file_data)
response = requests.get(urljoin(settings.API_URL, "/api/delete_old?path=" + full_path))
else:
response = requests.post(urljoin(settings.API_URL, "/api/upload_index"), data=file_data)
response = requests.get(urljoin(settings.API_URL, "/api/delete_old?path="))
# delete local
shutil.rmtree(full_path)

2
codecov.yml Normal file
View File

@@ -0,0 +1,2 @@
ignore:
- "*/tests/*"

26
docker-compose-local.yaml Normal file
View File

@@ -0,0 +1,26 @@
version: "3.9"
services:
frontend:
build: ./frontend
environment:
- VITE_API_HOST=http://localhost:7091
- VITE_API_STREAMING=$VITE_API_STREAMING
- VITE_EMBEDDINGS_NAME=$EMBEDDINGS_NAME
ports:
- "5173:5173"
redis:
image: redis:6-alpine
ports:
- 6379:6379
mongo:
image: mongo:6
ports:
- 27017:27017
volumes:
- mongodb_data_container:/data/db
volumes:
mongodb_data_container:

22
docker-compose-mock.yaml Normal file
View File

@@ -0,0 +1,22 @@
version: "3.9"
services:
frontend:
build: ./frontend
environment:
- VITE_API_HOST=http://localhost:7091
- VITE_API_STREAMING=$VITE_API_STREAMING
ports:
- "5173:5173"
depends_on:
- mock-backend
mock-backend:
build: ./mock-backend
ports:
- "7091:7091"
redis:
image: redis:6-alpine
ports:
- 6379:6379

View File

@@ -14,12 +14,12 @@ services:
backend:
build: ./application
environment:
- API_KEY=$OPENAI_API_KEY
- EMBEDDINGS_KEY=$OPENAI_API_KEY
- API_KEY=$API_KEY
- EMBEDDINGS_KEY=$API_KEY
- LLM_NAME=$LLM_NAME
- CELERY_BROKER_URL=redis://redis:6379/0
- CELERY_RESULT_BACKEND=redis://redis:6379/1
- MONGO_URI=mongodb://mongo:27017/docsgpt
- SELF_HOSTED_MODEL=$SELF_HOSTED_MODEL
ports:
- "7091:7091"
volumes:
@@ -34,8 +34,9 @@ services:
build: ./application
command: celery -A application.app.celery worker -l INFO
environment:
- API_KEY=$OPENAI_API_KEY
- EMBEDDINGS_KEY=$OPENAI_API_KEY
- API_KEY=$API_KEY
- EMBEDDINGS_KEY=$API_KEY
- LLM_NAME=$LLM_NAME
- CELERY_BROKER_URL=redis://redis:6379/0
- CELERY_RESULT_BACKEND=redis://redis:6379/1
- MONGO_URI=mongodb://mongo:27017/docsgpt

51
docs/README.md Normal file
View File

@@ -0,0 +1,51 @@
# nextra-docsgpt
## Setting Up Docs Folder of DocsGPT Locally
### 1. Clone the DocsGPT repository:
```bash
git clone https://github.com/arc53/DocsGPT.git
```
### 2. Navigate to the docs folder:
```bash
cd DocsGPT/docs
```
The docs folder contains the markdown files that make up the documentation. The majority of the files are in the pages directory. Some notable files in this folder include:
`index.mdx`: The main documentation file.
`_app.js`: This file is used to customize the default Next.js application shell.
`theme.config.jsx`: This file is for configuring the Nextra theme for the documentation.
### 3. Verify that you have Node.js and npm installed in your system. You can check by running:
```bash
node --version
npm --version
```
### 4. If not installed, download Node.js and npm from the respective official websites.
### 5. Once you have Node.js and npm running, proceed to install yarn - another package manager that helps to manage project dependencies:
```bash
npm install --global yarn
```
### 6. Install the project dependencies using yarn:
```bash
yarn install
```
### 7. After the successful installation of the project dependencies, start the local server:
```bash
yarn dev
```
- Now, you should be able to view the docs on your local environment by visiting `http://localhost:5000`. You can explore the different markdown files and make changes as you see fit.
- **Footnotes:** This guide assumes you have Node.js and npm installed. The guide involves running a local server using yarn, and viewing the documentation offline. If you encounter any issues, it may be worth verifying your Node.js and npm installations and whether you have installed yarn correctly.

9
docs/next.config.js Normal file
View File

@@ -0,0 +1,9 @@
const withNextra = require('nextra')({
theme: 'nextra-theme-docs',
themeConfig: './theme.config.jsx'
})
module.exports = withNextra()
// If you have other Next.js configurations, you can pass them as the parameter:
// module.exports = withNextra({ /* other next.js config */ })

5976
docs/package-lock.json generated Normal file

File diff suppressed because it is too large Load Diff

17
docs/package.json Normal file
View File

@@ -0,0 +1,17 @@
{
"scripts":{
"dev": "next dev",
"build": "next build",
"start": "next start"
},
"license": "MIT",
"dependencies": {
"@vercel/analytics": "^1.0.2",
"docsgpt": "^0.2.4",
"next": "^13.5.1",
"nextra": "^2.12.3",
"nextra-theme-docs": "^2.12.3",
"react": "^18.2.0",
"react-dom": "^18.2.0"
}
}

View File

@@ -0,0 +1,110 @@
# Self-hosting DocsGPT on Amazon Lightsail
Here's a step-by-step guide on how to set up an Amazon Lightsail instance to host DocsGPT.
## Configuring your instance
(If you know how to create a Lightsail instance, you can skip to the recommended configuration part by clicking [here](#connecting-to-your-newly-created-instance)).
### 1. Create an AWS Account:
If you haven't already, create or log in to your AWS account at https://lightsail.aws.amazon.com.
### 2. Create an Instance:
a. Click "Create Instance."
b. Select the "Instance location." In most cases, the default location works fine.
c. Choose "Linux/Unix" as the image and "Ubuntu 20.04 LTS" as the Operating System.
d. Configure the instance plan based on your requirements. A "1 GB, 1vCPU, 40GB SSD, and 2TB transfer" setup is recommended for most scenarios.
e. Give your instance a unique name and click "Create Instance."
PS: It may take a few minutes for the instance setup to complete.
### Connecting to Your newly created Instance
Your instance will be ready a few minutes after creation. To access it, open the instance and click "Connect using SSH."
#### Clone the DocsGPT Repository
A terminal window will pop up, and the first step will be to clone the DocsGPT Git repository:
`git clone https://github.com/arc53/DocsGPT.git`
#### Download the package information
Once it has finished cloning the repository, it is time to download the package information from all sources. To do so, simply enter the following command:
`sudo apt update`
#### Install Docker and Docker Compose
DocsGPT backend and worker use Python, Frontend is written on React and the whole application is containerized using Docker. To install Docker and Docker Compose, enter the following commands:
`sudo apt install docker.io`
And now install docker-compose:
`sudo apt install docker-compose`
#### Access the DocsGPT Folder
Enter the following command to access the folder in which the DocsGPT docker-compose file is present.
`cd DocsGPT/`
#### Prepare the Environment
Inside the DocsGPT folder create a `.env` file and copy the contents of `.env_sample` into it.
`nano .env`
Make sure your `.env` file looks like this:
```
OPENAI_API_KEY=(Your OpenAI API key)
VITE_API_STREAMING=true
SELF_HOSTED_MODEL=false
```
To save the file, press CTRL+X, then Y, and then ENTER.
Next, set the correct IP for the Backend by opening the docker-compose.yml file:
`nano docker-compose.yml`
And Change line 7 to: `VITE_API_HOST=http://localhost:7091`
to this `VITE_API_HOST=http://<your instance public IP>:7091`
This will allow the frontend to connect to the backend.
#### Running the Application
You're almost there! Now that all the necessary bits and pieces have been installed, it is time to run the application. To do so, use the following command:
`sudo docker-compose up -d`
Launching it for the first time will take a few minutes to download all the necessary dependencies and build.
Once this is done you can go ahead and close the terminal window.
#### Enabling Ports
a. Before you are able to access your live instance, you must first enable the port that it is using.
b. Open your Lightsail instance and head to "Networking".
c. Then click on "Add rule" under "IPv4 Firewall", enter `5173` as your port, and hit "Create".
Repeat the process for port `7091`.
#### Access your instance
Your instance is now available at your Public IP Address on port 5173. Enjoy using DocsGPT!
## Other Deployment Options
- [Deploy DocsGPT on Civo Compute Cloud](https://dev.to/rutamhere/deploying-docsgpt-on-civo-compute-c)
- [Deploy DocsGPT on DigitalOcean Droplet](https://dev.to/rutamhere/deploying-docsgpt-on-digitalocean-droplet-50ea)
- [Deploy DocsGPT on Kamatera Performance Cloud](https://dev.to/rutamhere/deploying-docsgpt-on-kamatera-performance-cloud-1bj)

View File

@@ -0,0 +1,128 @@
## Launching Web App
**Note**: Make sure you have Docker installed
**On macOS or Linux:**
Just run the following command:
```bash
./setup.sh
```
This command will install all the necessary dependencies and provide you with an option to use our LLM API, download the local model or use OpenAI.
If you prefer to follow manual steps, refer to this guide:
1. Open and download this repository with
```bash
git clone https://github.com/arc53/DocsGPT.git
```
2. Create a `.env` file in your root directory and set your `API_KEY` with your [OpenAI API key](https://platform.openai.com/account/api-keys). (optional in case you want to use OpenAI)
3. Run the following commands:
```bash
docker-compose build && docker-compose up
```
4. Navigate to http://localhost:5173/.
To stop, simply press **Ctrl + C**.
**For WINDOWS:**
To run the setup on Windows, you have two options: using the Windows Subsystem for Linux (WSL) or using Git Bash or Command Prompt.
**Option 1: Using Windows Subsystem for Linux (WSL):**
1. Install WSL if you haven't already. You can follow the official Microsoft documentation for installation: (https://learn.microsoft.com/en-us/windows/wsl/install).
2. After setting up WSL, open the WSL terminal.
3. Clone the repository and create the `.env` file:
```bash
git clone https://github.com/arc53/DocsGPT.git
cd DocsGPT
echo "API_KEY=Yourkey" > .env
echo "VITE_API_STREAMING=true" >> .env
```
4. Run the following command to start the setup with Docker Compose:
```bash
./run-with-docker-compose.sh
```
6. Open your web browser and navigate to http://localhost:5173/.
7. To stop the setup, just press **Ctrl + C** in the WSL terminal
**Option 2: Using Git Bash or Command Prompt (CMD):**
1. Install Git for Windows if you haven't already. Download it from the official website: (https://gitforwindows.org/).
2. Open Git Bash or Command Prompt.
3. Clone the repository and create the `.env` file:
```bash
git clone https://github.com/arc53/DocsGPT.git
cd DocsGPT
echo "API_KEY=Yourkey" > .env
echo "VITE_API_STREAMING=true" >> .env
```
4. Run the following command to start the setup with Docker Compose:
```bash
./run-with-docker-compose.sh
```
5. Open your web browser and navigate to http://localhost:5173/.
6. To stop the setup, just press **Ctrl + C** in the Git Bash or Command Prompt terminal.
These steps should help you set up and run the project on Windows using either WSL or Git Bash/Command Prompt.
**Important:** Ensure that Docker is installed and properly configured on your Windows system for these steps to work.
For WINDOWS:
To run the given setup on Windows, you can use the Windows Subsystem for Linux (WSL) or a Git Bash terminal to execute similar commands. Here are the steps adapted for Windows:
Option 1: Using Windows Subsystem for Linux (WSL):
1. Install WSL if you haven't already. You can follow the official Microsoft documentation for installation: (https://learn.microsoft.com/en-us/windows/wsl/install).
2. After setting up WSL, open the WSL terminal.
3. Clone the repository and create the `.env` file:
```bash
git clone https://github.com/arc53/DocsGPT.git
cd DocsGPT
echo "API_KEY=Yourkey" > .env
echo "VITE_API_STREAMING=true" >> .env
```
4. Run the following command to start the setup with Docker Compose:
```bash
./run-with-docker-compose.sh
```
5. Open your web browser and navigate to http://localhost:5173/.
6. To stop the setup, just press **Ctrl + C** in the WSL terminal.
Option 2: Using Git Bash or Command Prompt (CMD):
1. Install Git for Windows if you haven't already. You can download it from the official website: (https://gitforwindows.org/).
2. Open Git Bash or Command Prompt.
3. Clone the repository and create the `.env` file:
```bash
git clone https://github.com/arc53/DocsGPT.git
cd DocsGPT
echo "API_KEY=Yourkey" > .env
echo "VITE_API_STREAMING=true" >> .env
```
4. Run the following command to start the setup with Docker Compose:
```bash
./run-with-docker-compose.sh
```
5. Open your web browser and navigate to http://localhost:5173/.
6. To stop the setup, just press **Ctrl + C** in the Git Bash or Command Prompt terminal.
These steps should help you set up and run the project on Windows using either WSL or Git Bash/Command Prompt. Make sure you have Docker installed and properly configured on your Windows system for this to work.
### Chrome Extension
#### Installing the Chrome extension:
To enhance your DocsGPT experience, you can install the DocsGPT Chrome extension. Here's how:
1. In the DocsGPT GitHub repository, click on the **Code** button and select **Download ZIP**.
2. Unzip the downloaded file to a location you can easily access.
3. Open the Google Chrome browser and click on the three dots menu (upper right corner).
4. Select **More Tools** and then **Extensions**.
5. Turn on the **Developer mode** switch in the top right corner of the **Extensions page**.
6. Click on the **Load unpacked** button.
7. Select the **Chrome** folder where the DocsGPT files have been unzipped (docsgpt-main > extensions > chrome).
8. The extension should now be added to Google Chrome and can be managed on the Extensions page.
9. To disable or remove the extension, simply turn off the toggle switch on the extension card or click the **Remove** button.

View File

@@ -0,0 +1,254 @@
# Self-hosting DocsGPT on Railway
Here's a step-by-step guide on how to host DocsGPT on Railway App.
At first Clone and set up the project locally to run , test and Modify.
### 1. Clone and GitHub SetUp
a. Open Terminal (Windows Shell or Git bash(recommended)).
b. Type `git clone https://github.com/arc53/DocsGPT.git`
#### Download the package information
Once it has finished cloning the repository, it is time to download the package information from all sources. To do so, simply enter the following command:
`sudo apt update`
#### Install Docker and Docker Compose
DocsGPT backend and worker use Python, Frontend is written on React and the whole application is containerized using Docker. To install Docker and Docker Compose, enter the following commands:
`sudo apt install docker.io`
And now install docker-compose:
`sudo apt install docker-compose`
#### Access the DocsGPT Folder
Enter the following command to access the folder in which the DocsGPT docker-compose file is present.
`cd DocsGPT/`
#### Prepare the Environment
Inside the DocsGPT folder create a `.env` file and copy the contents of `.env_sample` into it.
`nano .env`
Make sure your `.env` file looks like this:
```
OPENAI_API_KEY=(Your OpenAI API key)
VITE_API_STREAMING=true
SELF_HOSTED_MODEL=false
```
To save the file, press CTRL+X, then Y, and then ENTER.
Next, set the correct IP for the Backend by opening the docker-compose.yml file:
`nano docker-compose.yml`
And Change line 7 to: `VITE_API_HOST=http://localhost:7091`
to this `VITE_API_HOST=http://<your instance public IP>:7091`
This will allow the frontend to connect to the backend.
#### Running the Application
You're almost there! Now that all the necessary bits and pieces have been installed, it is time to run the application. To do so, use the following command:
`sudo docker-compose up -d`
Launching it for the first time will take a few minutes to download all the necessary dependencies and build.
Once this is done you can go ahead and close the terminal window.
### 2. Pushing it to your own Repository
a. Create a Repository on your GitHub.
b. Open Terminal in the same directory of the Cloned project.
c. Type `git init`
d. `git add .`
e. `git commit -m "first-commit"`
f. `git remote add origin <your repository link>`
g. `git push git push --set-upstream origin master`
Your local files will now be pushed to your GitHub Account. :)
### 3. Create a Railway Account:
If you haven't already, create or log in to your railway account do it by visiting [Railway](https://railway.app/)
Signup via **GitHub** [Recommended].
### 4. Start New Project:
a. Open Railway app and Click on "Start New Project."
b. Choose any from the list of options available (Recommended "**Deploy from GitHub Repo**")
c. Choose the required Repository from your GitHub.
d. Configure and allow access to modify your GitHub content from the pop-up window.
e. Agree to all the terms and conditions.
PS: It may take a few minutes for the account setup to complete.
#### You will get A free trial of $5 (use it for trial and then purchase if satisfied and needed)
### 5. Connecting to Your newly Railway app with GitHub
a. Choose DocsGPT repo from the list of your GitHub repository that you want to deploy now.
b. Click on Deploy now.
![Three Tabs will be there](/Railway-selection.png)
c. Select Variables Tab.
d. Upload the env file here that you used for local setup.
e. Go to Settings Tab now.
f. Go to "Networking" and click on Generate Domain Name, to get the URL of your hosted project.
g. You can update the Root directory, build command, installation command as per need.
*[However recommended not the disturb these options and leave them as default if not that needed.]*
Your own DocsGPT is now available at the Generated domain URl. :)

View File

@@ -0,0 +1,14 @@
{
"Hosting-the-app": {
"title": "☁️ Hosting DocsGPT",
"href": "/Deploying/Hosting-the-app"
},
"Quickstart": {
"title": "⚡Quickstart",
"href": "/Deploying/Quickstart"
},
"Railway-Deploying": {
"title": "🚂Deploying on Railway",
"href": "/Deploying/Railway-Deploying"
}
}

View File

@@ -0,0 +1,229 @@
# API Endpoints Documentation
*Currently, the application provides the following main API endpoints:*
### 1. /api/answer
**Description:**
This endpoint is used to request answers to user-provided questions.
**Request:**
**Method**: `POST`
**Headers**: Content-Type should be set to `application/json; charset=utf-8`
**Request Body**: JSON object with the following fields:
* `question` — The user's question.
* `history` — (Optional) Previous conversation history.
* `api_key`— Your API key.
* `embeddings_key` — Your embeddings key.
* `active_docs` — The location of active documentation.
Here is a JavaScript Fetch Request example:
```js
// answer (POST http://127.0.0.1:5000/api/answer)
fetch("http://127.0.0.1:5000/api/answer", {
"method": "POST",
"headers": {
"Content-Type": "application/json; charset=utf-8"
},
"body": JSON.stringify({"question":"Hi","history":null,"api_key":"OPENAI_API_KEY","embeddings_key":"OPENAI_API_KEY",
"active_docs": "javascript/.project/ES2015/openai_text-embedding-ada-002/"})
})
.then((res) => res.text())
.then(console.log.bind(console))
```
**Response**
In response, you will get a JSON document containing the `answer`, `query` and `result`:
```json
{
"answer": "Hi there! How can I help you?\n",
"query": "Hi",
"result": "Hi there! How can I help you?\nSOURCES:"
}
```
### 2. /api/docs_check
**Description:**
This endpoint will make sure documentation is loaded on the server (just run it every time user is switching between libraries (documentations)).
**Request:**
**Method**: `POST`
**Headers**: Content-Type should be set to `application/json; charset=utf-8`
**Request Body**: JSON object with the field:
* `docs` — The location of the documentation:
```js
// docs_check (POST http://127.0.0.1:5000/api/docs_check)
fetch("http://127.0.0.1:5000/api/docs_check", {
"method": "POST",
"headers": {
"Content-Type": "application/json; charset=utf-8"
},
"body": JSON.stringify({"docs":"javascript/.project/ES2015/openai_text-embedding-ada-002/"})
})
.then((res) => res.text())
.then(console.log.bind(console))
```
**Response:**
In response, you will get a JSON document like this one indicating whether the documentation exists or not:
```json
{
"status": "exists"
}
```
### 3. /api/combine
**Description:**
This endpoint provides information about available vectors and their locations with a simple GET request.
**Request:**
**Method**: `GET`
**Response:**
Response will include:
* `date`
* `description`
* `docLink`
* `fullName`
* `language`
* `location` (local or docshub)
* `model`
* `name`
* `version`
Example of JSON in Docshub and local:
<img width="295" alt="image" src="https://user-images.githubusercontent.com/15183589/224714085-f09f51a4-7a9a-4efb-bd39-798029bb4273.png">
### 4. /api/upload
**Description:**
This endpoint is used to upload a file that needs to be trained, response is JSON with task ID, which can be used to check on task's progress.
**Request:**
**Method**: `POST`
**Request Body**: A multipart/form-data form with file upload and additional fields, including `user` and `name`.
HTML example:
```html
<form action="/api/upload" method="post" enctype="multipart/form-data" class="mt-2">
<input type="file" name="file" class="py-4" id="file-upload">
<input type="text" name="user" value="local" hidden>
<input type="text" name="name" placeholder="Name:">
<button type="submit" class="py-2 px-4 text-white bg-purple-30 rounded-md hover:bg-purple-30 focus:outline-none focus:ring-2 focus:ring-offset-2 focus:ring-purple-30">
Upload
</button>
</form>
```
**Response:**
JSON response with a status and a task ID that can be used to check the task's progress.
### 5. /api/task_status
**Description:**
This endpoint is used to get the status of a task (`task_id`) from `/api/upload`
**Request:**
**Method**: `GET`
**Query Parameter**: `task_id` (task ID to check)
**Sample JavaScript Fetch Request:**
```js
// Task status (Get http://127.0.0.1:5000/api/task_status)
fetch("http://localhost:5001/api/task_status?task_id=YOUR_TASK_ID", {
"method": "GET",
"headers": {
"Content-Type": "application/json; charset=utf-8"
},
})
.then((res) => res.text())
.then(console.log.bind(console))
```
**Response:**
There are two types of responses:
1. While the task is still running, the 'current' value will show progress from 0 to 100.
```json
{
"result": {
"current": 1
},
"status": "PROGRESS"
}
```
2. When task is completed:
```json
{
"result": {
"directory": "temp",
"filename": "install.rst",
"formats": [
".rst",
".md",
".pdf"
],
"name_job": "somename",
"user": "local"
},
"status": "SUCCESS"
}
```
### 6. /api/delete_old
**Description:**
This endpoint is used to delete old Vector Stores.
**Request:**
**Method**: `GET`
**Query Parameter**: `task_id`
**Sample JavaScript Fetch Request:**
```js
// delete_old (GET http://127.0.0.1:5000/api/delete_old)
fetch("http://localhost:5001/api/delete_old?task_id=YOUR_TASK_ID", {
"method": "GET",
"headers": {
"Content-Type": "application/json; charset=utf-8"
},
})
.then((res) => res.text())
.then(console.log.bind(console))
```
**Response:**
JSON response indicating the status of the operation:
```json
{ "status": "ok" }
```

View File

@@ -0,0 +1,6 @@
{
"API-docs": {
"title": "🗂️️ API-docs",
"href": "/Developing/API-docs"
}
}

View File

@@ -0,0 +1,44 @@
## Chatwoot Extension Setup Guide
### Step 1: Prepare and Start DocsGPT
- **Launch DocsGPT**: Follow the instructions in our [DocsGPT Wiki](https://github.com/arc53/DocsGPT/wiki) to start DocsGPT. Make sure to load your documentation.
### Step 2: Get Access Token from Chatwoot
- Go to Chatwoot.
- In your profile settings (located at the bottom left), scroll down and copy the **Access Token**.
### Step 3: Set Up Chatwoot Extension
- Navigate to `/extensions/chatwoot`.
- Copy the `.env_sample` file and create a new file named `.env`.
- Fill in the values in the `.env` file as follows:
```env
docsgpt_url=<Docsgpt_API_URL>
chatwoot_url=<Chatwoot_URL>
docsgpt_key=<OpenAI_API_Key or Other_LLM_Key>
chatwoot_token=<Token from Step 2>
```
### Step 4: Start the Extension
- Use the command `flask run` to start the extension.
### Step 5: Optional - Extra Validation
- In app.py, uncomment lines 12-13 and 71-75.
- Add the following lines to your .env file:
```account_id=(optional) 1
assignee_id=(optional) 1
```
These Chatwoot values help ensure you respond to the correct widget and handle questions assigned to a specific user.
### Stopping Bot Responses for Specific User or Session
- If you want the bot to stop responding to questions for a specific user or session, add a label `human-requested` in your conversation.
### Additional Notes
- For further details on training on other documentation, refer to our [wiki](https://github.com/arc53/DocsGPT/wiki/How-to-train-on-other-documentation).

View File

@@ -0,0 +1,10 @@
{
"Chatwoot-extension": {
"title": "💬️ Chatwoot Extension",
"href": "/Extensions/Chatwoot-extension"
},
"react-widget": {
"title": "🏗️ Widget setup",
"href": "/Extensions/react-widget"
}
}

View File

@@ -0,0 +1,47 @@
### Setting up the DocsGPT Widget in Your React Project
### Introduction:
The DocsGPT Widget is a powerful tool that allows you to integrate AI-powered documentation assistance into your web applications. This guide will walk you through the installation and usage of the DocsGPT Widget in your React project. Whether you're building a web app or a knowledge base, this widget can enhance your user experience.
### Installation
First, make sure you have Node.js and npm installed in your project. Then go to your project and install a new dependency: `npm install docsgpt`.
### Usage
In the file where you want to use the widget, import it and include the CSS file:
```js
import { DocsGPTWidget } from "docsgpt";
import "docsgpt/dist/style.css";
```
Now, you can use the widget in your component like this :
```jsx
<DocsGPTWidget
apiHost="https://your-docsgpt-api.com"
selectDocs="local/docs.zip"
apiKey=""
/>
```
DocsGPTWidget takes 3 **props**:
1. `apiHost` — The URL of your DocsGPT API.
2. `selectDocs` — The documentation source that you want to use for your widget (e.g. `default` or `local/docs1.zip`).
3. `apiKey` — Usually, it's empty.
### How to use DocsGPTWidget with [Nextra](https://nextra.site/) (Next.js + MDX)
Install your widget as described above and then go to your `pages/` folder and create a new file `_app.js` with the following content:
```js
import { DocsGPTWidget } from "docsgpt";
import "docsgpt/dist/style.css";
export default function MyApp({ Component, pageProps }) {
return (
<>
<Component {...pageProps} />
<DocsGPTWidget selectDocs="local/docsgpt-sep.zip/"/>
</>
)
}
```
For more information about React, refer to this [link here](https://react.dev/learn)

View File

@@ -0,0 +1,27 @@
# Customizing the Main Prompt
Customizing the main prompt for DocsGPT gives you the ability to tailor the AI's responses to your specific requirements. By modifying the prompt text, you can achieve more accurate and relevant answers. Here's how you can do it:
1. Navigate to `/application/prompts/combine_prompt.txt`.
2. Open the `combine_prompt.txt` file and modify the prompt text to suit your needs. You can experiment with different phrasings and structures to observe how the model responds. The main prompt serves as guidance to the AI model on how to generate responses.
## Example Prompt Modification
**Original Prompt:**
```markdown
You are a DocsGPT, friendly and helpful AI assistant by Arc53 that provides help with documents. You give thorough answers with code examples if possible.
Use the following pieces of context to help answer the users question. If it's not relevant to the question, provide friendly responses.
You have access to chat history, and can use it to help answer the question.
When using code examples, use the following format:
(code)
{summaries}
```
Feel free to customize the prompt to align it with your specific use case or the kind of responses you want from the AI. For example, you can focus on specific document types, industries, or topics to get more targeted results.
## Conclusion
Customizing the main prompt for DocsGPT allows you to tailor the AI's responses to your unique requirements. Whether you need in-depth explanations, code examples, or specific insights, you can achieve it by modifying the main prompt. Remember to experiment and fine-tune your prompts to get the best results.

View File

@@ -0,0 +1,63 @@
## How to train on other documentation
This AI can utilize any documentation, but it requires preparation for similarity search. Follow these steps to get your documentation ready:
**Step 1: Prepare Your Documentation**
![video-example-of-how-to-do-it](https://d3dg1063dc54p9.cloudfront.net/videos/how-to-vectorise.gif)
Start by going to `/scripts/` folder.
If you open this file, you will see that it uses RST files from the folder to create a `index.faiss` and `index.pkl`.
It currently uses OPENAI to create the vector store, so make sure your documentation is not too large. Using Pandas cost me around $3-$4.
You can typically find documentation on GitHub in the `docs/` folder for most open-source projects.
### 1. Find documentation in .rst/.md format and create a folder with it in your scripts directory.
- Name it `inputs/`.
- Put all your .rst/.md files in there.
- The search is recursive, so you don't need to flatten them.
If there are no .rst/.md files, convert whatever you find to a .txt file and feed it. (Don't forget to change the extension in the script).
### Step 2: Configure Your OpenAI API Key
1. Create a .env file in the scripts/ folder.
- Add your OpenAI API key inside: OPENAI_API_KEY=<your-api-key>.
### Step 3: Run the Ingestion Script
`python ingest.py ingest`
It will provide you with the estimated cost.
### Step 4: Move `index.faiss` and `index.pkl` generated in `scripts/output` to `application/` folder.
### Step 5: Run the Web App
Once you run it, it will use new context relevant to your documentation.Make sure you select default in the dropdown in the UI.
## Customization
You can learn more about options while running ingest.py by running:
- Make sure you select 'default' from the dropdown in the UI.
## Customization
You can learn more about options while running ingest.py by executing:
`python ingest.py --help`
| Options | |
|:--------------------------------:|:------------------------------------------------------------------------------------------------------------------------------:|
| **ingest** | Runs 'ingest' function, converting documentation to Faiss plus Index format |
| --dir TEXT | List of paths to directory for index creation. E.g. --dir inputs --dir inputs2 [default: inputs] |
| --file TEXT | File paths to use (Optional; overrides directory) E.g. --files inputs/1.md --files inputs/2.md |
| --recursive / --no-recursive | Whether to recursively search in subdirectories [default: recursive] |
| --limit INTEGER | Maximum number of files to read |
| --formats TEXT | List of required extensions (list with .) Currently supported: .rst, .md, .pdf, .docx, .csv, .epub, .html [default: .rst, .md] |
| --exclude / --no-exclude | Whether to exclude hidden files (dotfiles) [default: exclude] |
| -y, --yes | Whether to skip price confirmation |
| --sample / --no-sample | Whether to output sample of the first 5 split documents. [default: no-sample] |
| --token-check / --no-token-check | Whether to group small documents and split large. Improves semantics. [default: token-check] |
| --min_tokens INTEGER | Minimum number of tokens to not group. [default: 150] |
| --max_tokens INTEGER | Maximum number of tokens to not split. [default: 2000] |
| | |
| **convert** | Creates documentation in .md format from source code |
| --dir TEXT | Path to a directory with source code. E.g. --dir inputs [default: inputs] |
| --formats TEXT | Source code language from which to create documentation. Supports py, js and java. E.g. --formats py [default: py] |

View File

@@ -0,0 +1,48 @@
# Setting Up Local Language Models for Your App
Your app relies on two essential models: Embeddings and Text Generation. While OpenAI's default models work seamlessly, you have the flexibility to switch providers or even run the models locally.
## Step 1: Configure Environment Variables
Navigate to the `.env` file or set the following environment variables:
```env
LLM_NAME=<your Text Generation model>
API_KEY=<API key for Text Generation>
EMBEDDINGS_NAME=<LLM for Embeddings>
EMBEDDINGS_KEY=<API key for Embeddings>
VITE_API_STREAMING=<true or false>
```
You can omit the keys if users provide their own. Ensure you set `LLM_NAME` and `EMBEDDINGS_NAME`.
## Step 2: Choose Your Models
**Options for `LLM_NAME`:**
- openai ([More details](https://platform.openai.com/docs/models))
- anthropic ([More details](https://docs.anthropic.com/claude/reference/selecting-a-model))
- manifest ([More details](https://python.langchain.com/docs/integrations/llms/manifest))
- cohere ([More details](https://docs.cohere.com/docs/llmu))
- llama.cpp ([More details](https://python.langchain.com/docs/integrations/llms/llamacpp))
- huggingface (Arc53/DocsGPT-7B by default)
- sagemaker ([Mode details](https://aws.amazon.com/sagemaker/))
Note: for huggingface you can choose any model inside application/llm/huggingface.py or pass llm_name on init, loads
**Options for `EMBEDDINGS_NAME`:**
- openai_text-embedding-ada-002
- huggingface_sentence-transformers/all-mpnet-base-v2
- huggingface_hkunlp/instructor-large
- cohere_medium
If you want to be completely local, set `EMBEDDINGS_NAME` to `huggingface_sentence-transformers/all-mpnet-base-v2`.
For llama.cpp Download the required model and place it in the `models/` folder.
Alternatively, for local Llama setup, run `setup.sh` and choose option 1. The script handles the DocsGPT model addition.
## Step 3: Local Hosting for Privacy
If working with sensitive data, host everything locally by setting `LLM_NAME`, llama.cpp or huggingface, use any model available on Hugging Face, for llama.cpp you need to convert it into gguf format.
That's it! Your app is now configured for local and private hosting, ensuring optimal security for critical data.

View File

@@ -0,0 +1,21 @@
# Avoiding hallucinations
If your AI uses external knowledge and is not explicit enough, it is ok, because we try to make DocsGPT friendly.
But if you want to adjust it, here is a simple way:-
- Got to `application/prompts/chat_combine_prompt.txt`
- And change it to
```
You are a DocsGPT, friendly and helpful AI assistant by Arc53 that provides help with documents. You give thorough answers with code examples, if possible.
Write an answer for the question below based on the provided context.
If the context provides insufficient information, reply "I cannot answer".
You have access to chat history and can use it to help answer the question.
----------------
{summaries}
```

View File

@@ -0,0 +1,18 @@
{
"Customising-prompts": {
"title": "🏗️️ Customising Prompts",
"href": "/Guides/Customising-prompts"
},
"How-to-train-on-other-documentation": {
"title": "📥 Training on docs",
"href": "/Guides/How-to-train-on-other-documentation"
},
"How-to-use-different-LLM": {
"title": "⚙️️ How to use different LLM's",
"href": "/Guides/How-to-use-different-LLM"
},
"My-AI-answers-questions-using-external-knowledge": {
"title": "💭️ Avoiding hallucinations",
"href": "/Guides/My-AI-answers-questions-using-external-knowledge"
}
}

11
docs/pages/_app.js Normal file
View File

@@ -0,0 +1,11 @@
import { DocsGPTWidget } from "docsgpt";
import "docsgpt/dist/style.css";
export default function MyApp({ Component, pageProps }) {
return (
<>
<Component {...pageProps} />
<DocsGPTWidget selectDocs="local/docsgpt-sep.zip/"/>
</>
)
}

37
docs/pages/index.mdx Normal file
View File

@@ -0,0 +1,37 @@
---
title: 'Home'
---
import { Cards, Card } from 'nextra/components'
import deployingGuides from './Deploying/_meta.json';
import developingGuides from './Developing/_meta.json';
import extensionGuides from './Extensions/_meta.json';
import mainGuides from './Guides/_meta.json';
export const allGuides = {
...deployingGuides,
...developingGuides,
...extensionGuides,
...mainGuides,
};
### **DocsGPT 🦖**
DocsGPT 🦖 is an innovative open-source tool designed to simplify the retrieval of information from project documentation using advanced GPT models 🤖. Eliminate lengthy manual searches 🔍 and enhance your documentation experience with DocsGPT, and consider contributing to its AI-powered future 🚀.
![video-example-of-docs-gpt](https://d3dg1063dc54p9.cloudfront.net/videos/demov3.gif)
Try it yourself: [https://docsgpt.arc53.com/](https://docsgpt.arc53.com/)
<Cards
num={3}
children={Object.keys(allGuides).map((key, i) => (
<Card
key={i}
title={allGuides[key].title}
href={allGuides[key].href}
/>
))}
/>

Binary file not shown.

After

Width:  |  Height:  |  Size: 11 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 191 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 17 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 1.1 KiB

Some files were not shown because too many files have changed in this diff Show More