mirror of
https://github.com/TeamWiseFlow/wiseflow.git
synced 2025-01-23 02:20:20 +08:00
46 lines
1.7 KiB
Markdown
46 lines
1.7 KiB
Markdown
# Test Script Documentation
|
|
|
|
## Web Content Fetching and Parsing
|
|
|
|
[craw4ai_fetching.py](./craw4ai_fetching.py)
|
|
|
|
```
|
|
python craw4ai_fetching.py -S 'url1,url2...'
|
|
```
|
|
|
|
## HTML Content Parsing
|
|
|
|
[pre_process_test.py](./pre_process_test.py)
|
|
|
|
```
|
|
python pre_process_test.py -F 'json_file_path' -R 'record save path'
|
|
```
|
|
|
|
## Large Model Information Extraction Testing
|
|
|
|
[get_info_test.py](./get_info_test.py)
|
|
|
|
- To create focus point descriptions for test tasks, refer to [reports/wiseflow_report_v036_bigbrother666/task0/focus_point.json](./reports/wiseflow_report_v036_bigbrother666/task0/focus_point.json)
|
|
|
|
```
|
|
python get_info_test.py -D 'sample dir' -I 'include ap'
|
|
```
|
|
|
|
*-I whether to test LLM extraction of author and publish date*
|
|
|
|
# Result Submission and Sharing
|
|
|
|
Wiseflow is an open source project aiming to create an "information crawling tool for everyone" through collective contributions!
|
|
|
|
At this stage, **submitting test results is equivalent to submitting project code** - you'll be accepted as a contributor and may even be invited to participate in commercial projects!
|
|
|
|
Test results should be submitted to the [reports](./reports) directory. Create a subdirectory for each test named `{test_content}_{test_date}_{tester}`, for example:
|
|
|
|
```bash
|
|
mkdir -p reports/wiseflow_report_v036_bigbrother666
|
|
```
|
|
|
|
Please submit all test samples and the original output results of the program run, and create a README.md file in the directory to record the test content, test date, tester, test models, conclusions, statistical tables, etc.
|
|
|
|
Finally, edit the [reports/README.md](./reports/README.md) file, add the directory name of the test result to the index, so that others can view it.
|