Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Automated Test Generation for Tool-augmented LLMs #1469

Open
t1seungy opened this issue Mar 20, 2025 · 0 comments
Open

Automated Test Generation for Tool-augmented LLMs #1469

t1seungy opened this issue Mar 20, 2025 · 0 comments

Comments

@t1seungy
Copy link

Summary

A new test generation pipeline has been proposed to evaluate tool-augmented LLMs as conversational AI agents. This framework uses LLMs to generate diverse tests grounded on user-defined procedures, ensuring high coverage of possible conversations.

Implementation Guidance

  • Implement the test generation pipeline to evaluate LLMs in conversational AI scenarios.
  • Utilize the ALMITA dataset for evaluating AI agents in customer support and other domains.

Reference

Automated test generation to evaluate tool-augmented LLMs as conversational AI agents

Tags

  • LLM
  • Conversational AI
  • Testing

Assignee

@ComposioHQ

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant