Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

AI-Enhanced Testing Framework: New Features for Test Case Generation, Code Review and Documentationöst improvement #1473

Open
wants to merge 2 commits into
base: main
Choose a base branch
from
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
258 changes: 258 additions & 0 deletions meta_ai.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,258 @@
import asyncio
import json
import logging
import os
import re
import uuid
from pathlib import Path
from typing import Callable

import aiohttp.web
import pytest

from metagpt.const import DEFAULT_WORKSPACE_ROOT, TEST_DATA_PATH
from metagpt.context import Context as MetagptContext
from metagpt.llm import LLM
from metagpt.logs import logger
from metagpt.utils.git_repository import GitRepository
from metagpt.utils.project_repo import ProjectRepo
from tests.mock.mock_aiohttp import MockAioResponse
from tests.mock.mock_curl_cffi import MockCurlCffiResponse
from tests.mock.mock_httplib2 import MockHttplib2Response
from tests.mock.mock_llm import MockLLM

RSP_CACHE_NEW = {} # used globally for producing new and useful only response cache
ALLOW_OPENAI_API_CALL = int(
os.environ.get("ALLOW_OPENAI_API_CALL", 1)
) # NOTE: should change to default 0 (False) once mock is complete


@pytest.fixture(scope="session")
def rsp_cache():
rsp_cache_file_path = TEST_DATA_PATH / "rsp_cache.json" # read repo-provided
new_rsp_cache_file_path = TEST_DATA_PATH / "rsp_cache_new.json" # exporting a new copy
if os.path.exists(rsp_cache_file_path):
with open(rsp_cache_file_path, "r", encoding="utf-8") as f1:
rsp_cache_json = json.load(f1)
else:
rsp_cache_json = {}
yield rsp_cache_json
with open(rsp_cache_file_path, "w", encoding="utf-8") as f2:
json.dump(rsp_cache_json, f2, indent=4, ensure_ascii=False)
with open(new_rsp_cache_file_path, "w", encoding="utf-8") as f2:
json.dump(RSP_CACHE_NEW, f2, indent=4, ensure_ascii=False)


# AI-based test case generation
def generate_test_cases(context):
prompt = f"Generate test cases for the following context: {context}"
ai_generated_cases = LLM().ask(prompt)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It should be await LLM().aask(xx) and this function should also be async

return ai_generated_cases

@pytest.fixture(scope="session")
def auto_generated_tests():
context_info = "description of context"
generated_tests = generate_test_cases(context_info)
for test in generated_tests:
eval(test)


# AI-driven code review
def ai_review_code(code_snippet):
prompt = f"Review the following code for potential improvements: {code_snippet}"
suggestions = LLM().ask(prompt)
logger.info(f"AI Suggestions: {suggestions}")
return suggestions

@pytest.fixture(scope="function", autouse=True)
def llm_review(mocker, request):
if hasattr(request.node, "test_outcome") and request.node.test_outcome.passed:
ai_review_code(request.node.function)


# AI-powered mock data generation
async def ai_generate_mock_response(prompt):
response = await LLM().aask(prompt)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It should be await LLM().aask(xx) and this function should also be async

return response

@pytest.fixture
def aiohttp_mocker_with_ai(mocker):
async def mock_request(method, url, **kwargs):
prompt = f"Generate a mock {method} response for URL: {url}"
response = await ai_generate_mock_response(prompt)
return response

mocker.patch("aiohttp.ClientSession.request", mock_request)


# AI-driven error detection and correction
def ai_error_correction(log):
prompt = f"Detect and suggest fixes for the following error log: {log}"
fix_suggestions = LLM().ask(prompt)
logger.info(f"AI Suggested Fixes: {fix_suggestions}")
return fix_suggestions

@pytest.hookimpl(tryfirst=True, hookwrapper=True)
def pytest_runtest_makereport(item, call):
outcome = yield
rep = outcome.get_result()
if rep.when == "call" and rep.failed:
ai_error_correction(rep.longreprtext)


# AI-assisted documentation generation
def ai_generate_docs(context):
prompt = f"Generate documentation for the following context: {context}"
docs = LLM().ask(prompt)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It should be await LLM().aask(xx) and this function should also be async

return docs

@pytest.fixture(scope="session")
def generate_test_docs():
context_info = "Context of the test suite"
docs = ai_generate_docs(context_info)
with open('test_documentation.md', 'w') as f:
f.write(docs)


@pytest.fixture(scope="function", autouse=True)
def llm_mock(rsp_cache, mocker, request):
llm = MockLLM(allow_open_api_call=ALLOW_OPENAI_API_CALL)
llm.rsp_cache = rsp_cache
mocker.patch("metagpt.provider.base_llm.BaseLLM.aask", llm.aask)
mocker.patch("metagpt.provider.base_llm.BaseLLM.aask_batch", llm.aask_batch)
mocker.patch("metagpt.provider.openai_api.OpenAILLM.aask_code", llm.aask_code)
yield mocker
if hasattr(request.node, "test_outcome") and request.node.test_outcome.passed:
if llm.rsp_candidates:
for rsp_candidate in llm.rsp_candidates:
cand_key = list(rsp_candidate.keys())[0]
cand_value = list(rsp_candidate.values())[0]
if cand_key not in llm.rsp_cache:
logger.info(f"Added '{cand_key[:100]} ... -> {str(cand_value)[:20]} ...' to response cache")
llm.rsp_cache.update(rsp_candidate)
RSP_CACHE_NEW.update(rsp_candidate)


class Context:
def __init__(self):
self._llm_ui = None
self._llm_api = LLM()

@property
def llm_api(self):
# 1. Initialize llm with cache results
# 2. If query cached, return cached results
# 3. If no cached query, call llm_api and return results
# 4. If query cached, update cached results
return self._llm_api


@pytest.fixture(scope="package")
def llm_api():
logger.info("Setting up the test")
g_context = Context()

yield g_context.llm_api

logger.info("Tearing down the test")


@pytest.fixture
def proxy():
pattern = re.compile(
rb"(?P<method>[a-zA-Z]+) (?P<uri>(\w+://)?(?P<host>[^\s\'\"<>\[\]{}|/:]+)(:(?P<port>\d+))?[^\s\'\"<>\[\]{}|]*) "
)

async def pipe(reader, writer):
while not reader.at_eof():
writer.write(await reader.read(2048))
writer.close()
await writer.wait_closed()

async def handle_client(reader, writer):
data = await reader.readuntil(b"\r\n\r\n")
infos = pattern.match(data)
host, port = infos.group("host"), infos.group("port")
print(f"Proxy: {host}") # checking with capfd fixture
port = int(port) if port else 80
remote_reader, remote_writer = await asyncio.open_connection(host, port)
if data.startswith(b"CONNECT"):
writer.write(b"HTTP/1.1 200 Connection Established\r\n\r\n")
else:
remote_writer.write(data)
await asyncio.gather(pipe(reader, remote_writer), pipe(remote_reader, writer))

async def proxy_func():
server = await asyncio.start_server(handle_client, "127.0.0.1", 0)
return server, "http://{}:{}".format(*server.sockets[0].getsockname())

return proxy_func


# see https://github.com/Delgan/loguru/issues/59#issuecomment-466591978
@pytest.fixture
def loguru_caplog(caplog):
class PropogateHandler(logging.Handler):
def emit(self, record):
logging.getLogger(record.name).handle(record)

logger.add(PropogateHandler(), format="{message}")
yield caplog


@pytest.fixture(scope="function")
def context(request):
ctx = MetagptContext()
ctx.git_repo = GitRepository(local_path=DEFAULT_WORKSPACE_ROOT / f"unittest/{uuid.uuid4().hex}")
ctx.repo = ProjectRepo(ctx.git_repo)

# Destroy git repo at the end of the test session.
def fin():
if ctx.git_repo:
ctx.git_repo.delete_repository()

# Register the function for destroying the environment.
request.addfinalizer(fin)
return ctx



@pytest.fixture(scope="session", autouse=True)
def your_function_name():
# This function overrides the base class method but does not add any new behavior.
pass

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

maybe we could remove this function currently


@pytest.fixture(scope="function")
def new_filename(mocker):
# NOTE: Mock new filename to make reproducible llm aask, should consider changing after implementing requirement segmentation
mocker.patch("metagpt.utils.file_repository.FileRepository.new_filename", lambda: "20240101")
yield mocker


def _rsp_cache(name):
rsp_cache_file_path = TEST_DATA_PATH / f"{name}.json" # read repo-provided
if os.path.exists(rsp_cache_file_path):
with open(rsp_cache_file_path, "r") as f1:
rsp_cache_json = json.load(f1)
else:
rsp_cache_json = {}
yield rsp_cache_json
with open(rsp_cache_file_path, "w") as f2:
json.dump(rsp_cache_json, f2, indent=4, ensure_ascii=False)


@pytest.fixture(scope="session")
def rsp_cache_prompt():
return _rsp_cache("rsp_cache_prompt")


@pytest.fixture(scope="session")
def rsp_cache_code():
return _rsp_cache("rsp_cache_code")


@pytest.fixture(scope="session")
def rsp_cache_ctx():
return _rsp_cache("rsp_cache_ctx")

Loading