Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature Request] Implement KnowNO for HITL with structure output #1542

Open
1 of 2 tasks
lightaime opened this issue Feb 3, 2025 · 2 comments
Open
1 of 2 tasks

[Feature Request] Implement KnowNO for HITL with structure output #1542

lightaime opened this issue Feb 3, 2025 · 2 comments
Assignees
Labels
New Feature P1 Task with middle level priority
Milestone

Comments

@lightaime
Copy link
Member

Required prerequisites

Motivation

In dynamic and unfamiliar environments, large models and robots often face a common problem: making overly confident yet incorrect predictions. A team of researchers from Princeton University and Google DeepMind addressed this issue by introducing the KnowNo framework. This system helps robots recognize when they’re uncertain and allows them to ask for help from humans when necessary, using a concept called conformal prediction (CP).

Solution

No response

Alternatives

No response

Additional context

No response

@lightaime lightaime added the enhancement New feature or request label Feb 3, 2025
@lightaime lightaime added this to the Sprint 22 milestone Feb 3, 2025
@Wendong-Fan Wendong-Fan added New Feature call for contribution P0 Task with high level priority P1 Task with middle level priority and removed enhancement New feature or request P0 Task with high level priority call for contribution labels Feb 3, 2025
@Wendong-Fan
Copy link
Member

lead @MuggleJinx , support&review @X-TRON404 , @WHALEEYE

@MuggleJinx
Copy link
Collaborator

MuggleJinx commented Feb 18, 2025

Some thoughts about the implementations, I think the core idea of the paper is that:

  1. Leverage the ability of model (especially reasoning model) to give multiple potential next steps.
  2. Review the previous choices, (e.g., A, B, C, D), assign them with potential probabilities / credibilities.
  3. Compare with the user given threshold, if there are choices reach the threshold, take the step, otherwise let user review the choices and decide.

Some toy example I just came up:
agent.step_with_reasoning("Who is the best football player in the world? Tell me about his career.", choices=3, threshold=0.5 )

(Agent processing)
I think potential choices are A. Messi, B. Ronaldo, C. Neymar ....

Now I need to assign credibility to all choices:

A. Messi, 0.42
B. Ronaldo, 0.38
C. Neymar, 0.2

Since no choice is above 0.5, I should ask user to decide.

(call `ask_human_via_console`)

User input: I prefer Messi, plz continue the task:

(agent continues...)

What do you think? We can add one more argument in step() and implement this process internally.

@lightaime @Wendong-Fan

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
New Feature P1 Task with middle level priority
Projects
Status: No status
Development

No branches or pull requests

3 participants