Skip to content

Commit 485ce9d

Browse files
authoredMar 3, 2023
flake8 pipelines examples,ui,restapi (PaddlePaddle#5078)
1 parent b4f11f9 commit 485ce9d

14 files changed

+89
-129
lines changed
 

‎pipelines/examples/FAQ/dense_faq_example.py

+9-7
Original file line numberDiff line numberDiff line change
@@ -12,15 +12,17 @@
1212
# See the License for the specific language governing permissions and
1313
# limitations under the License.
1414

15-
### 城市百科知识智能问答系统
15+
# 城市百科知识智能问答系统
1616
import argparse
17-
import logging
1817
import os
1918

20-
import paddle
2119
from pipelines.document_stores import FAISSDocumentStore
22-
from pipelines.utils import convert_files_to_dicts, fetch_archive_from_http, print_documents
23-
from pipelines.nodes import ErnieRanker, DensePassageRetriever
20+
from pipelines.nodes import DensePassageRetriever, ErnieRanker
21+
from pipelines.utils import (
22+
convert_files_to_dicts,
23+
fetch_archive_from_http,
24+
print_documents,
25+
)
2426

2527
# yapf: disable
2628
parser = argparse.ArgumentParser()
@@ -82,10 +84,10 @@ def dense_faq_pipeline():
8284
# save index
8385
document_store.save(args.index_name)
8486

85-
### Ranker
87+
# Ranker
8688
ranker = ErnieRanker(model_name_or_path="rocketqa-zh-dureader-cross-encoder", use_gpu=use_gpu)
8789

88-
# ### Pipeline
90+
# Pipeline
8991
from pipelines import SemanticSearchPipeline
9092

9193
pipe = SemanticSearchPipeline(retriever, ranker)

‎pipelines/examples/document-intelligence/docprompt_example.py

+1-4
Original file line numberDiff line numberDiff line change
@@ -13,12 +13,9 @@
1313
# limitations under the License.
1414

1515
import argparse
16-
import logging
17-
import os
1816

19-
import paddle
20-
from pipelines.nodes import DocOCRProcessor, DocPrompter
2117
from pipelines import DocPipeline
18+
from pipelines.nodes import DocOCRProcessor, DocPrompter
2219

2320
# yapf: disable
2421
parser = argparse.ArgumentParser()

‎pipelines/examples/question-answering/dense_qa_example.py

+9-7
Original file line numberDiff line numberDiff line change
@@ -12,15 +12,17 @@
1212
# See the License for the specific language governing permissions and
1313
# limitations under the License.
1414

15-
### 城市百科知识智能问答系统
15+
# 城市百科知识智能问答系统
1616
import argparse
17-
import logging
1817
import os
1918

20-
import paddle
2119
from pipelines.document_stores import FAISSDocumentStore
22-
from pipelines.utils import convert_files_to_dicts, fetch_archive_from_http, print_answers
23-
from pipelines.nodes import ErnieReader, ErnieRanker, DensePassageRetriever
20+
from pipelines.nodes import DensePassageRetriever, ErnieRanker, ErnieReader
21+
from pipelines.utils import (
22+
convert_files_to_dicts,
23+
fetch_archive_from_http,
24+
print_answers,
25+
)
2426

2527
# yapf: disable
2628
parser = argparse.ArgumentParser()
@@ -82,14 +84,14 @@ def dense_qa_pipeline():
8284
# save index
8385
document_store.save(args.index_name)
8486

85-
### Ranker
87+
# Ranker
8688
ranker = ErnieRanker(model_name_or_path="rocketqa-zh-dureader-cross-encoder", use_gpu=use_gpu)
8789

8890
reader = ErnieReader(
8991
model_name_or_path="ernie-gram-zh-finetuned-dureader-robust", use_gpu=use_gpu, num_processes=1
9092
)
9193

92-
# ### Pipeline
94+
# Pipeline
9395
from pipelines import ExtractiveQAPipeline
9496

9597
pipe = ExtractiveQAPipeline(reader, ranker, retriever)

‎pipelines/examples/text_to_image/text_to_image_example.py

+3-5
Original file line numberDiff line numberDiff line change
@@ -12,12 +12,10 @@
1212
# See the License for the specific language governing permissions and
1313
# limitations under the License.
1414

15-
import os
1615
import argparse
1716

18-
import paddle
19-
from pipelines.nodes import ErnieTextToImageGenerator
2017
from pipelines import TextToImagePipeline
18+
from pipelines.nodes import ErnieTextToImageGenerator
2119

2220
# yapf: disable
2321
parser = argparse.ArgumentParser()
@@ -26,8 +24,7 @@
2624
parser.add_argument("--prompt_text", default='宁静的小镇', type=str, help="The prompt_text.")
2725
parser.add_argument("--output_dir", default='ernievilg_output', type=str, help="The output path.")
2826
parser.add_argument("--style", default='探索无限', type=str, help="The style text.")
29-
parser.add_argument("--size", default='1024*1024',
30-
choices=['1024*1024', '1024*1536', '1536*1024'], help="Size of the generation images")
27+
parser.add_argument("--size", default='1024*1024', choices=['1024*1024', '1024*1536', '1536*1024'], help="Size of the generation images")
3128
parser.add_argument("--topk", default=5, type=int, help="The top k images.")
3229
args = parser.parse_args()
3330
# yapf: enable
@@ -47,6 +44,7 @@ def text_to_image():
4744
}
4845
},
4946
)
47+
print(prediction)
5048
pipe.save_to_yaml("text_to_image.yaml")
5149

5250

‎pipelines/examples/unsupervised-question-answering/offline_question_answer_pairs_generation.py

+1-7
Original file line numberDiff line numberDiff line change
@@ -13,16 +13,10 @@
1313
# limitations under the License.
1414

1515
import argparse
16-
import logging
1716
import os
18-
from pprint import pprint
1917

20-
import paddle
2118
from pipelines.nodes import AnswerExtractor, QAFilter, QuestionGenerator
22-
from pipelines.nodes import ErnieRanker, DensePassageRetriever
23-
from pipelines.document_stores import FAISSDocumentStore
24-
from pipelines.utils import convert_files_to_dicts, fetch_archive_from_http, print_documents
25-
from pipelines.pipelines import QAGenerationPipeline, SemanticSearchPipeline
19+
from pipelines.pipelines import QAGenerationPipeline
2620

2721
# yapf: disable
2822
parser = argparse.ArgumentParser()

‎pipelines/examples/unsupervised-question-answering/unsupervised_question_answering_example.py

+9-6
Original file line numberDiff line numberDiff line change
@@ -13,16 +13,19 @@
1313
# limitations under the License.
1414

1515
import argparse
16-
import logging
1716
import os
1817
from pprint import pprint
1918

20-
import paddle
21-
from pipelines.nodes import AnswerExtractor, QAFilter, QuestionGenerator
22-
from pipelines.nodes import ErnieRanker, DensePassageRetriever
2319
from pipelines.document_stores import FAISSDocumentStore
24-
from pipelines.utils import convert_files_to_dicts, fetch_archive_from_http, print_documents
20+
from pipelines.nodes import (
21+
AnswerExtractor,
22+
DensePassageRetriever,
23+
ErnieRanker,
24+
QAFilter,
25+
QuestionGenerator,
26+
)
2527
from pipelines.pipelines import QAGenerationPipeline, SemanticSearchPipeline
28+
from pipelines.utils import convert_files_to_dicts, print_documents
2629

2730
# yapf: disable
2831
parser = argparse.ArgumentParser()
@@ -84,7 +87,7 @@ def dense_faq_pipeline():
8487
# save index
8588
document_store.save(args.index_name)
8689

87-
### Ranker
90+
# Ranker
8891
ranker = ErnieRanker(model_name_or_path="rocketqa-zh-dureader-cross-encoder", use_gpu=use_gpu)
8992

9093
pipe = SemanticSearchPipeline(retriever, ranker)

‎pipelines/pipelines/utils/export_utils.py

+3-3
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@
1616
import logging
1717
import pprint
1818
from collections import defaultdict
19-
from typing import Any, Dict, List, Optional
19+
from typing import Optional
2020

2121
import pandas as pd
2222

@@ -161,7 +161,7 @@ def export_answers_to_csv(agg_results: list, output_file):
161161
assert "query" in agg_results[0], f"Wrong format used for {agg_results[0]}"
162162
assert "answers" in agg_results[0], f"Wrong format used for {agg_results[0]}"
163163

164-
data = {} # type: Dict[str, List[Any]]
164+
data = {}
165165
data["query"] = []
166166
data["prediction"] = []
167167
data["prediction_rank"] = []
@@ -193,7 +193,7 @@ def convert_labels_to_squad(labels_file: str):
193193
for label in labels:
194194
labels_grouped_by_documents[label["document_id"]].append(label)
195195

196-
labels_in_squad_format = {"data": []} # type: Dict[str, Any]
196+
labels_in_squad_format = {"data": []}
197197
for document_id, labels in labels_grouped_by_documents.items():
198198
qas = []
199199
for label in labels:

‎pipelines/rest_api/application.py

+11-11
Original file line numberDiff line numberDiff line change
@@ -13,29 +13,29 @@
1313
# See the License for the specific language governing permissions and
1414
# limitations under the License.
1515

16-
import sys
1716
import logging
18-
19-
sys.path.append(".")
20-
21-
logging.basicConfig(format="%(asctime)s %(message)s", datefmt="%m/%d/%Y %I:%M:%S %p")
22-
logger = logging.getLogger(__name__)
23-
logging.getLogger("elasticsearch").setLevel(logging.WARNING)
24-
logging.getLogger("pipelines").setLevel(logging.INFO)
17+
import sys
2518

2619
import uvicorn
2720
from fastapi import FastAPI, HTTPException
28-
from fastapi.routing import APIRoute
2921
from fastapi.openapi.utils import get_openapi
22+
from fastapi.routing import APIRoute
3023
from starlette.middleware.cors import CORSMiddleware
3124

25+
# flake8: noqa
26+
sys.path.append(".")
27+
from rest_api.config import ROOT_PATH
3228
from rest_api.controller.errors.http_error import http_error_handler
33-
from rest_api.config import ROOT_PATH, PIPELINE_YAML_PATH
3429
from rest_api.controller.router import router as api_router
3530

31+
logging.basicConfig(format="%(asctime)s %(message)s", datefmt="%m/%d/%Y %I:%M:%S %p")
32+
logger = logging.getLogger(__name__)
33+
logging.getLogger("elasticsearch").setLevel(logging.WARNING)
34+
logging.getLogger("pipelines").setLevel(logging.INFO)
35+
3636
try:
3737
from pipelines import __version__ as pipelines_version
38-
except:
38+
except Exception:
3939
# For development
4040
pipelines_version = "0.0.0"
4141

‎pipelines/ui/webapp_docprompt_gradio.py

+8-10
Original file line numberDiff line numberDiff line change
@@ -14,19 +14,17 @@
1414
# See the License for the specific language governing permissions and
1515
# limitations under the License.
1616

17-
import os
18-
import json
17+
import argparse
1918
import base64
20-
from io import BytesIO
21-
from PIL import Image
2219
import traceback
23-
import argparse
20+
from io import BytesIO
2421

25-
import requests
26-
import numpy as np
27-
import gradio as gr
28-
import fitz
2922
import cv2
23+
import fitz
24+
import gradio as gr
25+
import numpy as np
26+
import requests
27+
from PIL import Image
3028

3129
fitz_tools = fitz.Tools()
3230

@@ -171,7 +169,7 @@ def read_content(file_path: str) -> str:
171169
padding-bottom: 2px !important;
172170
padding-left: 8px !important;
173171
padding-right: 8px !important;
174-
margin-top: 10px;
172+
margin-top: 10px;
175173
}
176174
.gradio-container .gr-button-primary {
177175
background: linear-gradient(180deg, #CDF9BE 0%, #AFF497 100%);

‎pipelines/ui/webapp_faq.py

+7-14
Original file line numberDiff line numberDiff line change
@@ -13,18 +13,16 @@
1313
# See the License for the specific language governing permissions and
1414
# limitations under the License.
1515

16+
import logging
1617
import os
1718
import sys
18-
import logging
19-
import pandas as pd
2019
from json import JSONDecodeError
2120
from pathlib import Path
21+
22+
import pandas as pd
2223
import streamlit as st
23-
from annotated_text import annotation
2424
from markdown import markdown
25-
26-
sys.path.append("ui")
27-
from utils import pipelines_is_ready, semantic_search, send_feedback, upload_doc, pipelines_version, get_backlink
25+
from utils import pipelines_is_ready, semantic_search, upload_doc
2826

2927
# Adjust to a question that you would like users to see in the search bar when they load the UI:
3028
DEFAULT_QUESTION_AT_STARTUP = os.getenv("DEFAULT_QUESTION_AT_STARTUP", "如何办理企业养老保险?")
@@ -58,7 +56,7 @@ def upload():
5856
for data_file in data_files:
5957
# Upload file
6058
if data_file and data_file.name not in st.session_state.upload_files["uploaded_files"]:
61-
raw_json = upload_doc(data_file)
59+
upload_doc(data_file)
6260
st.session_state.upload_files["uploaded_files"].append(data_file.name)
6361
# Save the uploaded files
6462
st.session_state.upload_files["uploaded_files"] = list(set(st.session_state.upload_files["uploaded_files"]))
@@ -115,16 +113,11 @@ def reset_results(*args):
115113
for data_file in st.session_state.upload_files["uploaded_files"]:
116114
st.sidebar.write(str(data_file) + "    ✅ ")
117115

118-
hs_version = ""
119-
try:
120-
hs_version = f" <small>(v{pipelines_version()})</small>"
121-
except Exception:
122-
pass
123116
# Load csv into pandas dataframe
124117
try:
125118
df = pd.read_csv(EVAL_LABELS, sep=";")
126119
except Exception:
127-
st.error(f"The eval file was not found.")
120+
st.error("The eval file was not found.")
128121
sys.exit(f"The eval file was not found under `{EVAL_LABELS}`.")
129122

130123
# Search bar
@@ -181,7 +174,7 @@ def reset_results(*args):
181174
st.session_state.results, st.session_state.raw_json = semantic_search(
182175
question, top_k_reader=top_k_reader, top_k_retriever=top_k_retriever
183176
)
184-
except JSONDecodeError as je:
177+
except JSONDecodeError:
185178
st.error("👓 &nbsp;&nbsp; An error occurred reading the results. Is the document store working?")
186179
return
187180
except Exception as e:

‎pipelines/ui/webapp_question_answering.py

+7-11
Original file line numberDiff line numberDiff line change
@@ -13,16 +13,17 @@
1313
# See the License for the specific language governing permissions and
1414
# limitations under the License.
1515

16+
import logging
1617
import os
1718
import sys
18-
import logging
19-
import pandas as pd
2019
from json import JSONDecodeError
2120
from pathlib import Path
21+
22+
import pandas as pd
2223
import streamlit as st
2324
from annotated_text import annotation
2425
from markdown import markdown
25-
from ui.utils import pipelines_is_ready, query, send_feedback, upload_doc, pipelines_version, get_backlink
26+
from ui.utils import get_backlink, pipelines_is_ready, query, upload_doc
2627

2728
# Adjust to a question that you would like users to see in the search bar when they load the UI:
2829
DEFAULT_QUESTION_AT_STARTUP = os.getenv("DEFAULT_QUESTION_AT_STARTUP", "中国的首都在哪里?")
@@ -54,7 +55,7 @@ def upload():
5455
for data_file in data_files:
5556
# Upload file
5657
if data_file and data_file.name not in st.session_state.upload_files["uploaded_files"]:
57-
raw_json = upload_doc(data_file)
58+
upload_doc(data_file)
5859
st.session_state.upload_files["uploaded_files"].append(data_file.name)
5960
# Save the uploaded files
6061
st.session_state.upload_files["uploaded_files"] = list(set(st.session_state.upload_files["uploaded_files"]))
@@ -109,7 +110,7 @@ def reset_results(*args):
109110
try:
110111
df = pd.read_csv(EVAL_LABELS, sep=";")
111112
except Exception:
112-
st.error(f"The eval file was not found.")
113+
st.error("The eval file was not found.")
113114
sys.exit(f"The eval file was not found under `{EVAL_LABELS}`.")
114115

115116
# File upload block
@@ -122,11 +123,6 @@ def reset_results(*args):
122123
st.sidebar.button("文件上传", on_click=upload)
123124
for data_file in st.session_state.upload_files["uploaded_files"]:
124125
st.sidebar.write(str(data_file) + " &nbsp;&nbsp; ✅ ")
125-
hs_version = ""
126-
try:
127-
hs_version = f" <small>(v{pipelines_version()})</small>"
128-
except Exception:
129-
pass
130126

131127
# Search bar
132128
question = st.text_input(
@@ -185,7 +181,7 @@ def reset_results(*args):
185181
st.session_state.results, st.session_state.raw_json = query(
186182
question, top_k_reader=top_k_reader, top_k_ranker=top_k_ranker, top_k_retriever=top_k_retriever
187183
)
188-
except JSONDecodeError as je:
184+
except JSONDecodeError:
189185
st.error("👓 &nbsp;&nbsp; An error occurred reading the results. Is the document store working?")
190186
return
191187
except Exception as e:

‎pipelines/ui/webapp_semantic_search.py

+7-16
Original file line numberDiff line numberDiff line change
@@ -13,20 +13,16 @@
1313
# See the License for the specific language governing permissions and
1414
# limitations under the License.
1515

16+
import logging
1617
import os
1718
import sys
18-
import logging
19-
import pandas as pd
2019
from json import JSONDecodeError
2120
from pathlib import Path
21+
22+
import pandas as pd
2223
import streamlit as st
23-
from annotated_text import annotation
2424
from markdown import markdown
25-
import socket
26-
27-
sys.path.append("ui")
28-
from utils import pipelines_is_ready, semantic_search, send_feedback, upload_doc, pipelines_version, get_backlink
29-
from utils import pipelines_files
25+
from utils import pipelines_files, pipelines_is_ready, semantic_search, upload_doc
3026

3127
# Adjust to a question that you would like users to see in the search bar when they load the UI:
3228
DEFAULT_QUESTION_AT_STARTUP = os.getenv("DEFAULT_QUESTION_AT_STARTUP", "衡量酒水的价格的因素有哪些?")
@@ -57,7 +53,7 @@ def upload():
5753
for data_file in data_files:
5854
# Upload file
5955
if data_file and data_file.name not in st.session_state.upload_files["uploaded_files"]:
60-
raw_json = upload_doc(data_file)
56+
upload_doc(data_file)
6157
st.session_state.upload_files["uploaded_files"].append(data_file.name)
6258
# Save the uploaded files
6359
st.session_state.upload_files["uploaded_files"] = list(set(st.session_state.upload_files["uploaded_files"]))
@@ -114,16 +110,11 @@ def reset_results(*args):
114110
for data_file in st.session_state.upload_files["uploaded_files"]:
115111
st.sidebar.write(str(data_file) + " &nbsp;&nbsp; ✅ ")
116112

117-
hs_version = ""
118-
try:
119-
hs_version = f" <small>(v{pipelines_version()})</small>"
120-
except Exception:
121-
pass
122113
# Load csv into pandas dataframe
123114
try:
124115
df = pd.read_csv(EVAL_LABELS, sep=";")
125116
except Exception:
126-
st.error(f"The eval file was not found.")
117+
st.error("The eval file was not found.")
127118
sys.exit(f"The eval file was not found under `{EVAL_LABELS}`.")
128119

129120
# Search bar
@@ -179,7 +170,7 @@ def reset_results(*args):
179170
st.session_state.results, st.session_state.raw_json = semantic_search(
180171
question, top_k_reader=top_k_reader, top_k_retriever=top_k_retriever
181172
)
182-
except JSONDecodeError as je:
173+
except JSONDecodeError:
183174
st.error("👓 &nbsp;&nbsp; An error occurred reading the results. Is the document store working?")
184175
return
185176
except Exception as e:

‎pipelines/ui/webapp_text_to_image.py

+1-4
Original file line numberDiff line numberDiff line change
@@ -12,13 +12,10 @@
1212
# See the License for the specific language governing permissions and
1313
# limitations under the License.
1414

15-
import re
16-
import os
1715
import argparse
1816

19-
from PIL import Image
20-
from utils import text_to_image_search
2117
import gradio as gr
18+
from utils import text_to_image_search
2219

2320
# yapf: disable
2421
parser = argparse.ArgumentParser()

‎pipelines/ui/webapp_unsupervised_question_answering.py

+13-24
Original file line numberDiff line numberDiff line change
@@ -13,26 +13,21 @@
1313
# See the License for the specific language governing permissions and
1414
# limitations under the License.
1515

16+
import logging
1617
import os
1718
import sys
18-
import logging
19-
import pandas as pd
2019
from json import JSONDecodeError
2120
from pathlib import Path
21+
22+
import pandas as pd
2223
import streamlit as st
23-
from annotated_text import annotation
2424
from markdown import markdown
25-
2625
from ui.utils import (
26+
file_upload_qa_generate,
27+
offline_ann,
2728
pipelines_is_ready,
2829
semantic_search,
29-
send_feedback,
30-
upload_doc,
31-
file_upload_qa_generate,
32-
pipelines_version,
33-
get_backlink,
3430
text_to_qa_pair_search,
35-
offline_ann,
3631
)
3732

3833
# Adjust to a question that you would like users to see in the search bar when they load the UI:
@@ -77,8 +72,7 @@ def upload():
7772
for data_file in data_files:
7873
# Upload file
7974
if data_file and data_file.name not in st.session_state.upload_files["uploaded_files"]:
80-
# raw_json = upload_doc(data_file)
81-
raw_json = file_upload_qa_generate(data_file)
75+
file_upload_qa_generate(data_file)
8276
st.session_state.upload_files["uploaded_files"].append(data_file.name)
8377
# Save the uploaded files
8478
st.session_state.upload_files["uploaded_files"] = list(set(st.session_state.upload_files["uploaded_files"]))
@@ -148,19 +142,14 @@ def reset_results_qag(*args):
148142
for data_file in st.session_state.upload_files["uploaded_files"]:
149143
st.sidebar.write(str(data_file) + " &nbsp;&nbsp; ✅ ")
150144

151-
hs_version = ""
152-
try:
153-
hs_version = f" <small>(v{pipelines_version()})</small>"
154-
except Exception:
155-
pass
156145
# Load csv into pandas dataframe
157146
try:
158147
df = pd.read_csv(EVAL_LABELS, sep=";")
159148
except Exception:
160-
st.error(f"The eval file was not found.")
149+
st.error("The eval file was not found.")
161150
sys.exit(f"The eval file was not found under `{EVAL_LABELS}`.")
162151

163-
## QA pairs generation
152+
# QA pairs generation
164153
# Search bar
165154
st.write("### 问答对生成:")
166155
context = st.text_input(
@@ -213,7 +202,7 @@ def reset_results_qag(*args):
213202
st.session_state.qag_results, st.session_state.qag_raw_json = text_to_qa_pair_search(
214203
context, is_filter=True if is_filter == "是" else False
215204
)
216-
except JSONDecodeError as je:
205+
except JSONDecodeError:
217206
st.error("👓 &nbsp;&nbsp; An error occurred reading the results. Is the document store working?")
218207
return
219208
except Exception as e:
@@ -229,9 +218,9 @@ def reset_results_qag(*args):
229218
for count, result in enumerate(st.session_state.qag_results):
230219
context = result["context"]
231220
synthetic_answer = result["synthetic_answer"]
232-
synthetic_answer_probability = result["synthetic_answer_probability"]
221+
# synthetic_answer_probability = result["synthetic_answer_probability"]
233222
synthetic_question = result["synthetic_question"]
234-
synthetic_question_probability = result["synthetic_question_probability"]
223+
# synthetic_question_probability = result["synthetic_question_probability"]
235224
st.write(
236225
markdown(context),
237226
unsafe_allow_html=True,
@@ -247,7 +236,7 @@ def reset_results_qag(*args):
247236

248237
st.write("___")
249238

250-
## QA search
239+
# QA search
251240
# Search bar
252241
st.write("### 问答检索:")
253242
question = st.text_input(
@@ -303,7 +292,7 @@ def reset_results_qag(*args):
303292
st.session_state.results, st.session_state.raw_json = semantic_search(
304293
question, top_k_reader=top_k_reader, top_k_retriever=top_k_retriever
305294
)
306-
except JSONDecodeError as je:
295+
except JSONDecodeError:
307296
st.error("👓 &nbsp;&nbsp; An error occurred reading the results. Is the document store working?")
308297
return
309298
except Exception as e:

0 commit comments

Comments
 (0)
Please sign in to comment.