update README.md
This commit is contained in:
11
README.md
11
README.md
@@ -44,13 +44,15 @@ python=3.10
|
||||
git clone https://github.com/OleehyO/TexTeller
|
||||
```
|
||||
|
||||
2. After [installing pytorch](https://pytorch.org/get-started/locally/#start-locally), install the project's dependencies:
|
||||
2. [Installing pytorch](https://pytorch.org/get-started/locally/#start-locally)
|
||||
|
||||
3. Install the project's dependencies:
|
||||
|
||||
```bash
|
||||
pip install -r requirements.txt
|
||||
```
|
||||
|
||||
3. Enter the `TexTeller/src` directory and run the following command in the terminal to start inference:
|
||||
4. Enter the `TexTeller/src` directory and run the following command in the terminal to start inference:
|
||||
|
||||
```bash
|
||||
python inference.py -img "/path/to/image.{jpg,png}"
|
||||
@@ -72,7 +74,10 @@ Go to the `TexTeller/src` directory and run the following command:
|
||||
Enter `http://localhost:8501` in a browser to view the web demo.
|
||||
|
||||
> [!TIP]
|
||||
> You can change the default configuration of `start_web.sh`, for example, to use GPU for inference (e.g. `USE_CUDA=True`) or to increase the number of beams (e.g. `NUM_BEAM=3`) to achieve higher accuracy
|
||||
> You can change the default configuration of `start_web.sh`, for example, to use GPU for inference (e.g. `USE_CUDA=True`) or to increase the number of beams (e.g. `NUM_BEAM=3`) to achieve higher accuracy.
|
||||
|
||||
> [!NOTE]
|
||||
> If you are Windows user, please run the `start_web.bat` file instead.
|
||||
|
||||
## 📡 API Usage
|
||||
|
||||
|
||||
@@ -44,13 +44,15 @@ python=3.10
|
||||
git clone https://github.com/OleehyO/TexTeller
|
||||
```
|
||||
|
||||
2. [安装pytorch](https://pytorch.org/get-started/locally/#start-locally)后,再安装本项目的依赖包:
|
||||
2. [安装pytorch](https://pytorch.org/get-started/locally/#start-locally)
|
||||
|
||||
3. 安装本项目的依赖包:
|
||||
|
||||
```bash
|
||||
pip install -r requirements.txt
|
||||
```
|
||||
|
||||
3. 进入`TexTeller/src`目录,在终端运行以下命令进行推理:
|
||||
4. 进入`TexTeller/src`目录,在终端运行以下命令进行推理:
|
||||
|
||||
```bash
|
||||
python inference.py -img "/path/to/image.{jpg,png}"
|
||||
@@ -102,6 +104,9 @@ python=3.10
|
||||
> [!TIP]
|
||||
> 你可以改变`start_web.sh`的默认配置, 例如使用GPU进行推理(e.g. `USE_CUDA=True`) 或者增加beams的数量(e.g. `NUM_BEAM=3`)来获得更高的精确度
|
||||
|
||||
> [!NOTE]
|
||||
> 对于Windows用户, 请运行 `start_web.bat`文件.
|
||||
|
||||
## 📡 API调用
|
||||
|
||||
我们使用[ray serve](https://github.com/ray-project/ray)来对外提供一个TexTeller的API接口,通过使用这个接口,你可以把TexTeller整合到自己的项目里。要想启动server,你需要先进入`TexTeller/src`目录然后运行以下命令:
|
||||
|
||||
11
src/start_web.bat
Normal file
11
src/start_web.bat
Normal file
@@ -0,0 +1,11 @@
|
||||
@echo off
|
||||
SETLOCAL ENABLEEXTENSIONS
|
||||
|
||||
set CHECKPOINT_DIR=default
|
||||
set TOKENIZER_DIR=default
|
||||
set USE_CUDA=False REM True or False (case-sensitive)
|
||||
set NUM_BEAM=1
|
||||
|
||||
streamlit run web.py
|
||||
|
||||
ENDLOCAL
|
||||
Reference in New Issue
Block a user