Skip to content

Commit 2b36c7d

Browse files
authored
fix starcoder example (#1080)
1 parent 9572f5d commit 2b36c7d

File tree

6 files changed

+11
-13
lines changed

6 files changed

+11
-13
lines changed

llm/inference/starcoder/README.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,9 @@ starcoder server for [llm-vscode](https://github.com/huggingface/llm-vscode) cus
2323
3. **Run application on port**
2424

2525
```shell
26-
python main.py --pretrained bigcode/starcoder2-7b
26+
python main.py --pretrained bigscience/starcoder2-7b
27+
# use mirror if you cannot access huggingface
28+
python main.py --pretrained AI-ModelScope/starcoder2-7b --mirror modelscope
2729
```
2830

2931
It may take a while to download the model checkpoint for the first time. Please change the mirror address by `--mirror` argument if it takes too long.
@@ -35,7 +37,7 @@ starcoder server for [llm-vscode](https://github.com/huggingface/llm-vscode) cus
3537
1) Select `Skip TLS verification for insecure connections`.
3638
![llm-vscode-settings](./assets/llm-vscode-TLS-verification.png)
3739

38-
2) Change `Llm:ConfigTemplate` to `Custom`.
40+
2) Change `Llm:Backend` to `tgi` and `Llm:ConfigTemplate` to `Custom`.
3941
![llm-vscode-settings](./assets/llm-vscode-config-template.png)
4042

4143
3) Select `Edit in settings.json` in `Llm:Url`.
15.4 KB
Loading
Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,5 @@
11
from mindnlp.transformers import GenerationConfig
22
from mindnlp.transformers import Pipeline, pipeline
3-
import mindspore
4-
53

64
class GeneratorBase:
75
def generate(self, query: str, parameters: dict) -> str:
@@ -12,12 +10,12 @@ def __call__(self, query: str, parameters: dict = None) -> str:
1210

1311

1412
class StarCoder(GeneratorBase):
15-
def __init__(self, pretrained: str, mirror: str = 'modelscope'):
13+
def __init__(self, pretrained: str, mirror: str = 'huggingface'):
1614
self.pretrained: str = pretrained
1715
self.mirror: str = mirror
1816
self.pipe: Pipeline = pipeline(
1917
"text-generation", model=pretrained, mirror=mirror)
20-
self.generation_config = GenerationConfig.from_pretrained(pretrained)
18+
self.generation_config = GenerationConfig.from_pretrained(pretrained, mirror=mirror)
2119
self.generation_config.pad_token_id = self.pipe.tokenizer.eos_token_id
2220

2321
def generate(self, query: str, parameters: dict) -> str:
@@ -27,4 +25,4 @@ def generate(self, query: str, parameters: dict) -> str:
2725
})
2826
json_response: dict = self.pipe(query, generation_config=config)[0]
2927
generated_text: str = json_response['generated_text']
30-
return generated_text
28+
return generated_text

llm/inference/starcoder/main.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -18,9 +18,7 @@ async def api(request: Request):
1818
json_request: dict = await request.json()
1919
inputs: str = json_request['inputs']
2020
parameters: dict = json_request['parameters']
21-
logger.info(f'{request.client.host}:{request.client.port} inputs = {json.dumps(inputs)}')
2221
generated_text: str = generator.generate(inputs, parameters)
23-
logger.info(f'{request.client.host}:{request.client.port} generated_text = {json.dumps(generated_text)}')
2422
return {
2523
"generated_text": generated_text.replace(inputs, ""),
2624
"status": 200
Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,3 @@
1-
uvicorn~=0.22.0
2-
fastapi~=0.95.1
3-
tokenizers~=0.15.0
1+
uvicorn
2+
fastapi
3+
mindnlp>=0.3.1

llm/inference/starcoder/util.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,6 @@ def get_parser() -> argparse.ArgumentParser:
1717
parser = argparse.ArgumentParser()
1818
parser.add_argument('--port', type=int, default=8000)
1919
parser.add_argument('--host', type=str, default='0.0.0.0')
20-
parser.add_argument('--mirror', type=str, default='modelscope')
20+
parser.add_argument('--mirror', type=str, default='huggingface')
2121
parser.add_argument('--pretrained', type=str, required=True)
2222
return parser

0 commit comments

Comments
 (0)