Add paper abstract to model card
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,10 +1,11 @@
|
|
| 1 |
---
|
| 2 |
-
license: cc-by-nc-4.0
|
| 3 |
datasets:
|
| 4 |
- Salesforce/APIGen-MT-5k
|
| 5 |
- Salesforce/xlam-function-calling-60k
|
| 6 |
language:
|
| 7 |
- en
|
|
|
|
|
|
|
| 8 |
pipeline_tag: text-generation
|
| 9 |
tags:
|
| 10 |
- function-calling
|
|
@@ -14,7 +15,6 @@ tags:
|
|
| 14 |
- qwen
|
| 15 |
- pytorch
|
| 16 |
- LLaMA-factory
|
| 17 |
-
library_name: transformers
|
| 18 |
---
|
| 19 |
|
| 20 |
<p align="center">
|
|
@@ -30,6 +30,10 @@ library_name: transformers
|
|
| 30 |
</p>
|
| 31 |
<hr>
|
| 32 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
# Welcome to the xLAM-2 Model Family!
|
| 34 |
|
| 35 |
[Large Action Models (LAMs)](https://blog.salesforceairesearch.com/large-action-models/) are advanced language models designed to enhance decision-making by translating user intentions into executable actions. As the **brains of AI agents**, LAMs autonomously plan and execute tasks to achieve specific goals, making them invaluable for automating workflows across diverse domains.
|
|
@@ -304,5 +308,4 @@ Additionally, please check our other awesome related works regarding xLAM series
|
|
| 304 |
journal={arXiv preprint arXiv:2402.15506},
|
| 305 |
year={2024}
|
| 306 |
}
|
| 307 |
-
```
|
| 308 |
-
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
datasets:
|
| 3 |
- Salesforce/APIGen-MT-5k
|
| 4 |
- Salesforce/xlam-function-calling-60k
|
| 5 |
language:
|
| 6 |
- en
|
| 7 |
+
library_name: transformers
|
| 8 |
+
license: cc-by-nc-4.0
|
| 9 |
pipeline_tag: text-generation
|
| 10 |
tags:
|
| 11 |
- function-calling
|
|
|
|
| 15 |
- qwen
|
| 16 |
- pytorch
|
| 17 |
- LLaMA-factory
|
|
|
|
| 18 |
---
|
| 19 |
|
| 20 |
<p align="center">
|
|
|
|
| 30 |
</p>
|
| 31 |
<hr>
|
| 32 |
|
| 33 |
+
## Paper Abstract
|
| 34 |
+
|
| 35 |
+
Training effective AI agents for multi-turn interactions requires high-quality data that captures realistic human-agent dynamics, yet such data is scarce and expensive to collect manually. We introduce APIGen-MT, a two-phase framework that generates verifiable and diverse multi-turn agent data. In the first phase, our agentic pipeline produces detailed task blueprints with ground-truth actions, leveraging a committee of LLM reviewers and iterative feedback loops. These blueprints are then transformed into complete interaction trajectories through simulated human-agent interplay. We train a family of models -- the xLAM-2-fc-r series with sizes ranging from 1B to 70B parameters. Our models outperform frontier models such as GPT-4o and Claude 3.5 on $\tau$-bench and BFCL benchmarks, with the smaller models surpassing their larger counterparts, particularly in multi-turn settings, while maintaining superior consistency across multiple trials. Comprehensive experiments demonstrate that our verified blueprint-to-details approach yields high-quality training data, enabling the development of more reliable, efficient, and capable agents. We open-source 5K synthetic data trajectories and the trained xLAM-2-fc-r models to advance research in AI agents. Models at this https URL Dataset at this https URL and Website at this https URL
|
| 36 |
+
|
| 37 |
# Welcome to the xLAM-2 Model Family!
|
| 38 |
|
| 39 |
[Large Action Models (LAMs)](https://blog.salesforceairesearch.com/large-action-models/) are advanced language models designed to enhance decision-making by translating user intentions into executable actions. As the **brains of AI agents**, LAMs autonomously plan and execute tasks to achieve specific goals, making them invaluable for automating workflows across diverse domains.
|
|
|
|
| 308 |
journal={arXiv preprint arXiv:2402.15506},
|
| 309 |
year={2024}
|
| 310 |
}
|
| 311 |
+
```
|
|
|