venkat-srinivasan-nexusflow commited on
Commit
c649912
•
1 Parent(s): 2ed01e6

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +77 -0
README.md ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Athene-V2-Agent: Surpassing GPT-4o for Tool Use And Agentic Usecases
2
+ <p align="center">
3
+ <a href="https://huggingface.co/Nexusflow" target="_blank">Nexusflow HF</a> - <a href="https://discord.gg/HDSVmNAs3y" target="_blank">Nexusflow Discord</a>
4
+ </p>
5
+
6
+ <p align="center" width="100%">
7
+ <a><img src="agent.png" alt="NexusRaven" style="width: 40%; min-width: 300px; display: block; margin: auto;"></a>
8
+ </p>
9
+
10
+ ## Introducing Athene-V2-Agent
11
+ Athene-V2-Agent is an open-source Agent LLM that surpasses the state-of-the-art in function calling and agentic capabilities.
12
+
13
+ 💪 **Versatile Agent Capability**: Athene-V2-Agent is an agent model, capable of operating in environments with deeply nested dependencies with the environment. It is capable of reasoning and doing planning for trajectories with many tool calls necessary to answer a single query.
14
+
15
+ 📊 **Performance Highlights**: Athene-V2-Agent surpasses GPT-4o in single FC tasks by 18% in function calling success rates, and by 17% in Agentic success rates.
16
+
17
+ 🔧 **Generalization to the Unseen**: Athene-V2-Agent has never been trained on the functions or agentic settings used in evaluation.
18
+
19
+ ## Athene-V2-Agent Model Usage
20
+
21
+ ### OpenAI-Compatible FC
22
+
23
+ Athene-V2-Agent is usable in any OpenAI API-compatible environment using our VLLM docker image. This should be a simple "drop-in" replacement to any agentic or tool-use setting with our VLLM docker image.
24
+
25
+ ```
26
+ docker run --name athene-v2-agent \
27
+ --runtime nvidia --gpus '"device=0,1,2,3,4,5,6,7"' \
28
+ -v ~/.cache/huggingface:/root/.cache/huggingface \
29
+ --env "HUGGING_FACE_HUB_TOKEN=<secret>" \
30
+ -p <port>:8000 \
31
+ --ipc=host \
32
+ ghcr.io/nexusflowai/athene-v2-vllm:latest \
33
+ --model Nexusflow/Athene-V2-Agent \
34
+ --dtype=auto \
35
+ --tensor-parallel-size=8 \
36
+ --enable-auto-tool-choice \
37
+ --tool-call-parser Athene-V2-Agent
38
+ ```
39
+
40
+ You can now submit any OpenAI-Compatible tool-use requests to the model by hitting the VLLM endpoint. Athene-V2-Agent will be able to issue tool calls that you can execute and return results for.
41
+
42
+ ### Examples
43
+
44
+ An example Weather agent for this can be found here: [Link](example/vllm_v2_weather_agent.py#L186-L193). This example includes handling Athene for queries that are answerable and not answerable by the current tools.
45
+
46
+ An example extraction and RAG-Agent can be found here: [Link](example/vllm_v2_extraction_agent.py#L270-L284). This example includes handling RAG-based queries with a wikipedia tool.
47
+
48
+ ### Prompting Tricks
49
+
50
+ 1. When giving docstrings to Athene-V2-Agent, please provide well-indented, detailed, and well-written docstrings as this can help accuracy.
51
+ 2. We strongly recommend using the docker image to interact with Athene-V2-Agent.
52
+ 4. We strongly recommend to set sampling to False when prompting Athene-V2-Agent.
53
+ 5. We strongly recommend a zero temperature.
54
+
55
+ When handling irrelevant user queries, users have noticed that specifying a "no-op" function with arguments work best. For example, something like this might work:
56
+ ```python
57
+ {
58
+ "type": "function",
59
+ "function" : {
60
+ "name": "no_relevant_function",
61
+ "description": "Call this when no other provided function can be called to answer the user query.",
62
+ "parameters": {
63
+ "type": "object",
64
+ "properties": {
65
+ "user_query_span": {
66
+ "type": "string",
67
+ "description": "The part of the user_query that cannot be answered by any other function calls."
68
+ }
69
+ },
70
+ "required": ["user_query_span"]
71
+ }
72
+ }
73
+ }
74
+ ```
75
+
76
+ ## Contact
77
+ Please join our [Discord Channel](https://discord.gg/HDSVmNAs3y) to reach out for any issues and comments!