OliveSensorAPI/demo
Anooyman de0674ccf7
Update main code (#2)
* update rag/src/data_processing.py

* Add files via upload

allow user to load embedding & rerank models from cache

* Add files via upload

embedding_path = os.path.join(model_dir, 'embedding_model')  
rerank_path = os.path.join(model_dir, 'rerank_model')

* 测试push dev

测试push dev

* Add files via upload

两个母亲多轮对话数据集合并、清理和去重之后,得到 2439 条多轮对话数据(每条有6-8轮对话)。

* optimize deduplicate.py

Add time print information
save duplicate dataset as well
remove print(content)

* add base model qlora fintuning config file: internlm2_7b_base_qlora_e10_M_1e4_32_64.py

* add full finetune code from internlm2

* other 2 configs for base model

* update cli_internlm2.py

 three methods to load model

1. download model in openxlab
2. download model in modelscope
3. offline model

* create upload_modelscope.py

* add base model and update personal contributions

* add README.md for Emollm_Scientist

* Create README_internlm2_7b_base_qlora.md

InternLM2 7B Base QLoRA 微调指南

* [DOC]EmoLLM_Scientist微调指南

* [DOC]EmoLLM_Scientist微调指南

* [DOC]EmoLLM_Scientist微调指南

* [DOC]EmoLLM_Scientist微调指南

* [DOC]EmoLLM_Scientist微调指南

* [DOC]EmoLLM_Scientist微调指南

* update

* [DOC]README_scientist.md

* delete config

* format update

* upload xlab

* add README_Model_Uploading.md and images

* modelscope model upload

* Modify Recent Updates

* update daddy-like Boy-Friend EmoLLM

* update model uploading with openxlab

* update model uploading with openxlab

---------

Co-authored-by: zealot52099 <songyan5209@163.com>
Co-authored-by: xzw <62385492+aJupyter@users.noreply.github.com>
Co-authored-by: zealot52099 <67356208+zealot52099@users.noreply.github.com>
Co-authored-by: Bryce Wang <90940753+brycewang2018@users.noreply.github.com>
Co-authored-by: HongCheng <kwchenghong@gmail.com>
2024-03-24 11:51:19 +08:00
..
cli_internlm2_scientist.py Update main code (#2) 2024-03-24 11:51:19 +08:00
cli_internlm2.py Update main code (#2) 2024-03-24 11:51:19 +08:00
cli_qwen.py feat: add web_internlm2 and upload s.t. scripts 2024-01-25 19:02:24 +08:00
README_EN.md README files translation 2024-03-03 19:24:55 +09:00
README.md README files translation 2024-03-03 19:24:55 +09:00
requirements_qwen.txt feat: finetune Qwen and demo 2024-01-21 19:12:03 +08:00
web_qwen.py feat: add web_internlm2 and upload s.t. scripts 2024-01-25 19:02:24 +08:00

Deploying Guide for EmoLLM

Local Deployment

  • Clone repo
git clone https://github.com/aJupyter/EmoLLM.git
  • Install dependencies
pip install -r requirements.txt
  • Download the model

  • cli_demo

python ./demo/cli_internlm2.py
  • web_demo
python ./app.py

If deploying on a server, you need to configure local port mapping.

Deploy on OpenXLab

  • Log in to OpenXLab and create a Gradio application

Login OpenXLab

  • Select configurations and create the project

config

  • Wait for the build and startup

wait a minutes

  • Try your own project

enjoy