Add paper abstract to model card

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +7 -4
README.md CHANGED
@@ -1,10 +1,11 @@
1
  ---
2
- license: cc-by-nc-4.0
3
  datasets:
4
  - Salesforce/APIGen-MT-5k
5
  - Salesforce/xlam-function-calling-60k
6
  language:
7
  - en
 
 
8
  pipeline_tag: text-generation
9
  tags:
10
  - function-calling
@@ -14,7 +15,6 @@ tags:
14
  - qwen
15
  - pytorch
16
  - LLaMA-factory
17
- library_name: transformers
18
  ---
19
 
20
  <p align="center">
@@ -30,6 +30,10 @@ library_name: transformers
30
  </p>
31
  <hr>
32
 
 
 
 
 
33
  # Welcome to the xLAM-2 Model Family!
34
 
35
  [Large Action Models (LAMs)](https://blog.salesforceairesearch.com/large-action-models/) are advanced language models designed to enhance decision-making by translating user intentions into executable actions. As the **brains of AI agents**, LAMs autonomously plan and execute tasks to achieve specific goals, making them invaluable for automating workflows across diverse domains.
@@ -304,5 +308,4 @@ Additionally, please check our other awesome related works regarding xLAM series
304
  journal={arXiv preprint arXiv:2402.15506},
305
  year={2024}
306
  }
307
- ```
308
-
 
1
  ---
 
2
  datasets:
3
  - Salesforce/APIGen-MT-5k
4
  - Salesforce/xlam-function-calling-60k
5
  language:
6
  - en
7
+ library_name: transformers
8
+ license: cc-by-nc-4.0
9
  pipeline_tag: text-generation
10
  tags:
11
  - function-calling
 
15
  - qwen
16
  - pytorch
17
  - LLaMA-factory
 
18
  ---
19
 
20
  <p align="center">
 
30
  </p>
31
  <hr>
32
 
33
+ ## Paper Abstract
34
+
35
+ Training effective AI agents for multi-turn interactions requires high-quality data that captures realistic human-agent dynamics, yet such data is scarce and expensive to collect manually. We introduce APIGen-MT, a two-phase framework that generates verifiable and diverse multi-turn agent data. In the first phase, our agentic pipeline produces detailed task blueprints with ground-truth actions, leveraging a committee of LLM reviewers and iterative feedback loops. These blueprints are then transformed into complete interaction trajectories through simulated human-agent interplay. We train a family of models -- the xLAM-2-fc-r series with sizes ranging from 1B to 70B parameters. Our models outperform frontier models such as GPT-4o and Claude 3.5 on $\tau$-bench and BFCL benchmarks, with the smaller models surpassing their larger counterparts, particularly in multi-turn settings, while maintaining superior consistency across multiple trials. Comprehensive experiments demonstrate that our verified blueprint-to-details approach yields high-quality training data, enabling the development of more reliable, efficient, and capable agents. We open-source 5K synthetic data trajectories and the trained xLAM-2-fc-r models to advance research in AI agents. Models at this https URL Dataset at this https URL and Website at this https URL
36
+
37
  # Welcome to the xLAM-2 Model Family!
38
 
39
  [Large Action Models (LAMs)](https://blog.salesforceairesearch.com/large-action-models/) are advanced language models designed to enhance decision-making by translating user intentions into executable actions. As the **brains of AI agents**, LAMs autonomously plan and execute tasks to achieve specific goals, making them invaluable for automating workflows across diverse domains.
 
308
  journal={arXiv preprint arXiv:2402.15506},
309
  year={2024}
310
  }
311
+ ```