Skip to content

Commit

Permalink
rename opengpt by rungpt
Browse files Browse the repository at this point in the history
  • Loading branch information
huangkaipeng4399 committed Aug 15, 2023
1 parent 13e6209 commit 171fabf
Show file tree
Hide file tree
Showing 92 changed files with 772 additions and 530 deletions.
22 changes: 11 additions & 11 deletions .github/workflows/docker-release.yml
Original file line number Diff line number Diff line change
Expand Up @@ -34,29 +34,29 @@ jobs:
echo "BUILD_DATE=$(date -u +'%Y-%m-%dT%H:%M:%SZ')" >> $GITHUB_ENV
echo "BUILD_TARGET=clip_executor" >> $GITHUB_ENV
VERSION=$(sed -n '/^__version__ = /p' ./open_gpt/__init__.py | cut -d \' -f2)
VERSION=$(sed -n '/^__version__ = /p' ./rungpt/__init__.py | cut -d \' -f2)
V_VERSION=v${VERSION}
MINOR_VERSION=${VERSION%.*}
MAJOR_VERSION=${MINOR_VERSION%.*}
if [[ "${{ github.event.inputs.triggered_by }}" == "CD" ]]; then
# on every CD release
echo "TAG_ALIAS=\
jinaai/open_gpt:master" \
jinaai/rungpt:master" \
>> $GITHUB_ENV
elif [[ "${{ github.event.inputs.triggered_by }}" == "TAG" ]]; then
# on every tag release
echo "TAG_ALIAS=\
jinaai/open_gpt:latest, \
jinaai/open_gpt:v${VERSION}, \
jinaai/open_gpt:v${MINOR_VERSION} \
jinaai/rungpt:latest, \
jinaai/rungpt:v${VERSION}, \
jinaai/rungpt:v${MINOR_VERSION} \
" >> $GITHUB_ENV
elif [[ "${{ github.event.inputs.triggered_by }}" == "MANUAL" ]]; then
# on every manual release
echo "TAG_ALIAS=\
jinaai/open_gpt:v${VERSION} \
jinaai/rungpt:v${VERSION} \
" >> $GITHUB_ENV
else
echo "Bad triggered_by: ${{ github.event.inputs.triggered_by }}!"
Expand Down Expand Up @@ -86,7 +86,7 @@ jobs:
with:
file: Dockerfiles/Dockerfile
platforms: linux/amd64
cache-from: type=registry,ref=jinaai/open_gpt:latest
cache-from: type=registry,ref=jinaai/rungpt:latest
cache-to: type=inline
push: true
tags: ${{env.TAG_ALIAS}}
Expand All @@ -101,10 +101,10 @@ jobs:
with:
file: Dockerfiles/gateway.Dockerfile
platforms: linux/amd64
cache-from: type=registry,ref=jinaai/open_gpt_gateway:latest
cache-from: type=registry,ref=jinaai/run_gpt_gateway:latest
cache-to: type=inline
push: true
tags: jinaai/open_gpt_gateway:v${{env.VERSION}}, jinaai/open_gpt_gateway:latest
tags: jinaai/run_gpt_gateway:v${{env.VERSION}}, jinaai/run_gpt_gateway:latest
build-args: |
BUILD_DATE=${{env.BUILD_DATE}}
VERSION=${{env.VERSION}}
Expand All @@ -116,10 +116,10 @@ jobs:
with:
file: Dockerfiles/executor.Dockerfile
platforms: linux/amd64
cache-from: type=registry,ref=jinaai/open_gpt_executor:latest
cache-from: type=registry,ref=jinaai/run_gpt_executor:latest
cache-to: type=inline
push: true
tags: jinaai/open_gpt_executor:v${{env.VERSION}}, jinaai/open_gpt_executor:latest
tags: jinaai/run_gpt_executor:v${{env.VERSION}}, jinaai/run_gpt_executor:latest
build-args: |
BUILD_DATE=${{env.BUILD_DATE}}
VERSION=${{env.VERSION}}
Expand Down
10 changes: 5 additions & 5 deletions .github/workflows/docs.yml
Original file line number Diff line number Diff line change
Expand Up @@ -9,12 +9,12 @@ jobs:
steps:
- uses: actions/checkout@v3
with:
path: opengpt
path: rungpt
- uses: actions/checkout@v3
with:
repository: numb3r3/opengpt.github.io.git
repository: numb3r3/rungpt.github.io.git
ref: 'main'
path: ./opengpt.github.io
path: ./rungpt.github.io
token: ${{ secrets.GH_TEST_TOKEN }}
- uses: actions/setup-python@v2
with:
Expand All @@ -23,5 +23,5 @@ jobs:
pip install pillow cairosvg
sudo apt-get install -y libcairo2-dev libfreetype6-dev libffi-dev libjpeg-dev libpng-dev libz-dev
pip install mkdocs-material mkdocs-material-extensions mkdocs-redirects --upgrade
mkdocs gh-deploy --config-file ../opengpt/mkdocs.yml --force
working-directory: ./opengpt.github.io
mkdocs gh-deploy --config-file ../rungpt/mkdocs.yml --force
working-directory: ./rungpt.github.io
2 changes: 1 addition & 1 deletion .github/workflows/force-release.yml
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@ jobs:
echo "VCS_REF=$VCS_REF" >> $GITHUB_ENV
echo "Will build $VCS_REF"
VERSION=$(sed -n '/^__version__ = /p' ./open_gpt/__init__.py | cut -d \' -f2)
VERSION=$(sed -n '/^__version__ = /p' ./run_gpt/__init__.py | cut -d \' -f2)
echo "VERSION=$VERSION" >> $GITHUB_ENV
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/release.yml
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ jobs:
echo "VCS_REF=$VCS_REF" >> $GITHUB_ENV
echo "Will build $VCS_REF"
VERSION=$(sed -n '/^__version__ = /p' ./open_gpt/__init__.py | cut -d \' -f2)
VERSION=$(sed -n '/^__version__ = /p' ./run_gpt/__init__.py | cut -d \' -f2)
echo "VERSION=$VERSION" >> $GITHUB_ENV
Expand Down
10 changes: 5 additions & 5 deletions .pre-commit-config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@ repos:
rev: 4.0.1
hooks:
- id: flake8
exclude: ^(.git|__pycache__|docs/source/conf.py|old|build|dist|tests|open_gpt/resources/)
exclude: ^(.git|__pycache__|docs/source/conf.py|old|build|dist|tests|run_gpt/resources/)
args:
- --max-complexity=10
- --max-line-length=127
Expand All @@ -12,8 +12,8 @@ repos:
# rev: v1.5.8
# hooks:
# - id: darglint
# files: open_gpt/
# exclude: ^(docs/|open_gpt/resources/)
# files: run_gpt/
# exclude: ^(docs/|run_gpt/resources/)
# args:
# - --message-template={path}:{line} {msg_id} {msg}
# - -s=sphinx
Expand All @@ -24,15 +24,15 @@ repos:
hooks:
- id: pydocstyle
files: client/
exclude: ^(docs/|open_gpt/resources/)
exclude: ^(docs/|run_gpt/resources/)
args:
- --select=D101,D102,D103
- repo: https://github.com/ambv/black
rev: 22.3.0
hooks:
- id: black
types: [python]
exclude: ^(docs/|open_gpt/resources/)
exclude: ^(docs/|run_gpt/resources/)
args:
- -S
- repo: https://github.com/asottile/blacken-docs
Expand Down
6 changes: 3 additions & 3 deletions Dockerfiles/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -8,9 +8,9 @@ FROM mosaicml/pytorch:${TORCH_VERSION}_cu${CUDA_VERSION}-python3.10-ubuntu20.04
ENV DEBIAN_FRONTEND=noninteractive LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8

# copy will almost always invalid the cache
COPY . /open_gpt/
WORKDIR /open_gpt
COPY . /run_gpt/
WORKDIR /run_gpt

RUN python3 -m pip install -e .

ENTRYPOINT ["opengpt"]
ENTRYPOINT ["rungpt"]
4 changes: 2 additions & 2 deletions MODEL_ZOO.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
# Model Zoo in OpenGPT
# Model Zoo in RunGPT

OpenGPT supports the following models out of the box:
RunGPT supports the following models out of the box:

- LLM (Large Language Model)

Expand Down
40 changes: 20 additions & 20 deletions README.md
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
# ☄️ OpenGPT
# ☄️ RunGPT

<p align="center">
<a href="https://github.com/jina-ai/opengpt"><img src="https://github.com/jina-ai/opengpt/blob/main/.github/images/logo.png" alt="OpenGPT: An open-source cloud-native large-scale multimodal model serving framework" width="300px"></a>
<a href="https://github.com/jina-ai/rungpt"><img src="https://github.com/jina-ai/rungpt/blob/main/.github/images/logo.png" alt="rungpt: An open-source cloud-native large-scale multimodal model serving framework" width="300px"></a>
<br>
</p>

Expand All @@ -11,10 +11,10 @@

![](https://img.shields.io/badge/Made%20with-JinaAI-blueviolet?style=flat)
[![PyPI](https://img.shields.io/pypi/v/open_gpt_torch)](https://pypi.org/project/open_gpt_torch/)
[![PyPI - License](https://img.shields.io/pypi/l/open_gpt_torch)](https://pypi.org/project/open_gpt_torch/)
[![PyPI](https://img.shields.io/pypi/v/run_gpt_torch)](https://pypi.org/project/run_gpt_torch/)
[![PyPI - License](https://img.shields.io/pypi/l/run_gpt_torch)](https://pypi.org/project/run_gpt_torch/)

**OpenGPT** is an open-source _cloud-native_ large-scale **_multimodal models_** (LMMs) serving framework.
**RunGPT** is an open-source _cloud-native_ large-scale **_multimodal models_** (LMMs) serving framework.
It is designed to simplify the deployment and management of large language models, on a distributed cluster of GPUs.
We aim to make it a one-stop solution for a centralized and accessible place to gather techniques for optimizing large-scale multimodal models and make them easy to use for everyone.

Expand All @@ -30,7 +30,7 @@ We aim to make it a one-stop solution for a centralized and accessible place to

## Features

OpenGPT provides the following features to make it easy to deploy and serve **large multi-modal models** (LMMs) at scale:
RunGPT provides the following features to make it easy to deploy and serve **large multi-modal models** (LMMs) at scale:

- Support for multi-modal models on top of large language models
- Scalable architecture for handling high traffic loads
Expand All @@ -41,13 +41,13 @@ OpenGPT provides the following features to make it easy to deploy and serve **la

## Updates

- **2023-05-12**: 🎉We have released the first version `v0.0.1` of OpenGPT. You can install it with `pip install open_gpt_torch`.
- **2023-05-12**: 🎉We have released the first version `v0.0.1` of RunGPT. You can install it with `pip install run_gpt_torch`.

## Supported Models

<details>

OpenGPT supports the following models out of the box:
RunGPT supports the following models out of the box:

- LLM (Large Language Model)

Expand All @@ -69,7 +69,7 @@ For more details about the supported models, please see the [Model Zoo](./MODEL_

## Roadmap

You can view our roadmap with features that are planned, started, and completed on the [Roadmap discussion](https://github.com/jina-ai/opengpt/discussions/categories/roadmap) category.
You can view our roadmap with features that are planned, started, and completed on the [Roadmap discussion](https://github.com/jina-ai/rungpt/discussions/categories/roadmap) category.

## Get Started

Expand All @@ -78,15 +78,15 @@ You can view our roadmap with features that are planned, started, and completed
Install the package with `pip`:

```bash
pip install open_gpt_torch
pip install run_gpt_torch
```

### Quickstart

```python
import open_gpt
import run_gpt

model = open_gpt.create_model(
model = run_gpt.create_model(
'stabilityai/stablelm-tuned-alpha-3b', device='cuda', precision='fp16'
)

Expand Down Expand Up @@ -117,7 +117,7 @@ We use the [stabilityai/stablelm-tuned-alpha-3b](https://huggingface.co/stabilit
In most cases of large model serving, the model cannot fit into a single GPU. To solve this problem, we also provide a `device_map` option (supported by `accecleate` package) to automatically partition the model and distribute it across multiple GPUs:

```python
model = open_gpt.create_model(
model = run_gpt.create_model(
'stabilityai/stablelm-tuned-alpha-3b', precision='fp16', device_map='balanced'
)
```
Expand All @@ -128,24 +128,24 @@ In the above example, `device_map="balanced"` evenly split the model on all avai
> The `device_map` option is supported by the [accelerate](https://github.com/huggingface/accelerate) package.

See [examples on how to use opengpt with different models.](./examples) 🔥
See [examples on how to use rungpt with different models.](./examples) 🔥


## Build a model serving in one line

To do so, you can use the `serve` command:

```bash
opengpt serve stabilityai/stablelm-tuned-alpha-3b --precision fp16 --device_map balanced
rungpt serve stabilityai/stablelm-tuned-alpha-3b --precision fp16 --device_map balanced
```

💡 **Tip**: you can inspect the available options with `opengpt serve --help`.
💡 **Tip**: you can inspect the available options with `rungpt serve --help`.

This will start a gRPC and HTTP server listening on port `51000` and `52000` respectively.
Once the server is ready, as shown below:
<details>
<summary>Click to expand</summary>
<img src="https://github.com/jina-ai/opengpt/blob/main/.github/images/serve_ready.png" width="600px">
<img src="https://github.com/jina-ai/rungpt/blob/main/.github/images/serve_ready.png" width="600px">
</details>

You can then send requests to the server:
Expand Down Expand Up @@ -173,7 +173,7 @@ response = requests.post(
What's more, we also provide a [Python client](https://github.com/jina-ai/inference-client/) (`inference-client`) for you to easily interact with the server:

```python
from open_gpt import Client
from run_gpt import Client

client = Client()

Expand Down Expand Up @@ -206,7 +206,7 @@ To do so, you can use `deploy` command:
using predefined executor

```bash
opengpt deploy stabilityai/stablelm-tuned-alpha-3b --precision fp16 --device_map balanced --cloud jina --replicas 1
rungpt deploy stabilityai/stablelm-tuned-alpha-3b --precision fp16 --device_map balanced --cloud jina --replicas 1
```

It will give you a HTTP url and a gRPC url by default:
Expand All @@ -226,4 +226,4 @@ We welcome contributions from the community! To contribute, please submit a pull

## License

OpenGPT is licensed under the Apache License, Version 2.0. See LICENSE for the full license text.
Rungpt is licensed under the Apache License, Version 2.0. See LICENSE for the full license text.
10 changes: 5 additions & 5 deletions docs/docs/index.md
Original file line number Diff line number Diff line change
@@ -1,18 +1,18 @@
# Quick start

`opengpt` is an open-source _cloud-native_ large-scale **_multimodal models_** (LMMs) serving framework.
`rungpt` is an open-source _cloud-native_ large-scale **_multimodal models_** (LMMs) serving framework.
It is designed to simplify the deployment and management of large language models, on a distributed cluster of GPUs.
We aim to make it a one-stop solution for a centralized and accessible place to gather techniques for optimizing large-scale multimodal models and make them easy to use for everyone.


## Installation and setup

To use `opengpt`, install it with `pip`:
To use `rungpt`, install it with `pip`:

<div class="termy">

```shell
$ pip install open_gpt_torch
$ pip install run_gpt_torch
```

</div>
Expand All @@ -25,9 +25,9 @@ We use the [stabilityai/stablelm-tuned-alpha-3b](https://huggingface.co/stabilit
<div class="termy">

```python
import open_gpt
import run_gpt

model = open_gpt.create_model(
model = run_gpt.create_model(
'stabilityai/stablelm-tuned-alpha-3b', device='cuda', precision='fp16'
)

Expand Down
8 changes: 4 additions & 4 deletions docs/overrides/home.html
Original file line number Diff line number Diff line change
Expand Up @@ -23,9 +23,9 @@
<div class="tx-landing__hero">
<div class="tx-landing__hero_text">
<h1>Run LLM faster and easier. Use any cloud.</h1>
<p>opengpt allows you to run your large-scale multi-modal models on any cloud, with a single command.</p>
<p>rungpt allows you to run your large-scale multi-modal models on any cloud, with a single command.</p>

<p>opengpt is open-source, self-hosted, and supports all major cloud providers,
<p>rungpt is open-source, self-hosted, and supports all major cloud providers,
including AWS, GCP, and Azure.</p>

<a href="/docs" class="md-button md-button--primary">
Expand Down Expand Up @@ -158,13 +158,13 @@ <h2>Open-source and self-hosted
d="M12 1.5A2.5 2.5 0 0 1 14.5 4 2.5 2.5 0 0 1 12 6.5 2.5 2.5 0 0 1 9.5 4 2.5 2.5 0 0 1 12 1.5M15.87 5C18 5 20 7 20 9c2.7 0 2.7 4 0 4H4c-2.7 0-2.7-4 0-4 0-2 2-4 4.13-4 .44 1.73 2.01 3 3.87 3 1.86 0 3.43-1.27 3.87-3M5 15h3l1 7H7l-2-7m5 0h4l-1 7h-2l-1-7m6 0h3l-2 7h-2l1-7Z"></path></svg></span>
</h2>
<p class="tx-landing__bottom_cta_text">
Getting started with opengpt's <a href="https://github.com/jina-ai/opengpt">open source</a> tool is
Getting started with rungpt's <a href="https://github.com/jina-ai/rungpt">open source</a> tool is
just a mater of:
</p>
<div class="tx-landing__bottom_cta_terminal">
<div class="tx-landing__bottom_cta_terminal_row">
<span class="tx-landing__bottom_cta_terminal_row_prefix">$</span>
pip install "open_gpt_torch"
pip install "run_gpt_torch"
</div>
</div>
<p class="tx-landing__bottom_cta_text">
Expand Down
4 changes: 2 additions & 2 deletions docs/overrides/main.html
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
{% extends "base.html" %}

{% block announce %}
Like opengpt? Give us a <img alt="" style="display: inline-block; height: 1.125em; margin-bottom: -2px"
Like rungpt? Give us a <img alt="" style="display: inline-block; height: 1.125em; margin-bottom: -2px"
src="https://cdnjs.cloudflare.com/ajax/libs/twemoji/14.0.2/svg/2b50.svg">
on <img style="display: inline-block; height: 1.15em; margin-bottom: -4px; margin-right: -2px"
src="{{ 'assets/images/github-logo.png' | url }}"/> <a href="https://github.com/jina-ai/opengpt" target="_blank">
src="{{ 'assets/images/github-logo.png' | url }}"/> <a href="https://github.com/jina-ai/rungpt" target="_blank">
GitHub</a><span style="color: var(--md-code-hl-punctuation-color)">!</span>
{% endblock %}
Loading

0 comments on commit 171fabf

Please sign in to comment.