HSC-Bench
Task definition, QoS metrics, composition datasets, model families, and results for workflow-level service composition.
Service composition converts complex user requirements into a service sequence, DAG, or executable workflow. Inputs include the user need, service library, functional constraints, input/output compatibility, and QoS constraints. The output should satisfy both functional requirements and non-functional optimization objectives.
| Dataset | Domain | QoS | Requirement Type | Description |
|---|---|---|---|---|
| QWS | Web Service | RT, TP, Availability, Reliability | Simulated requirements | Classic QoS service composition dataset. |
| WS-Dream | Web Service | Response Time, Throughput | QoS prediction oriented | Can support recommendation, composition, and QoS prediction; functional semantics are weaker. |
| HSC | AI Model Service | Metadata + QoS | Real-inspired workflows | Provides AI service workflows from Hugging Face model services. |
| HSC+ | AI Model Service | Response time, waiting time, reliability, successability | Generated realistic requirements and workflows | Core dataset for the benchmark. |
Genetic algorithm baseline for QoS-aware service composition.
CodeEvolutionary optimization method for multi-objective composition.
CodeWhale optimization variant for QoS composition search.
CodeSwarm intelligence baseline for service composition optimization.
CodeGenetic algorithm variant for service dependency and QoS constraints.
CodePopulation-based service composition optimization baseline.
CodeParticle swarm optimization method for QoS-aware composition.
CodeGraph neural network and pointer network based workflow generation model.
CodeReinforcement learning formulation for sequential composition decisions.
CodeSequence generation baseline for executable service workflows.
CodeAgentic service generation pipeline from natural-language need to service plan.
CodeAgent that translates workflow plans into executable code artifacts.
CodeTotal execution latency or path-level response time, typically aggregated along the workflow path.
Total service cost. If real cost is unavailable, the benchmark should define a simulated cost setting.
Workflow throughput, often determined by the bottleneck service.
Probability that the composed service is available, commonly multiplied across component services.
Probability of successful execution for the workflow.
Weighted normalized QoS score. The page should always document normalization and weights.
| Model | Dataset | Type | Utility ↑ | RT ↓ | Cost ↓ | Throughput ↑ | Availability ↑ | Reliability ↑ | Code | Official | Unified Protocol |
|---|---|---|---|---|---|---|---|---|---|---|---|
| GNNPN-SC | HSC+ | Learning-based | TBD | TBD | TBD | TBD | TBD | TBD | Link | Planned | Yes |
| SDFGA | HSC+ | Optimization-based | TBD | TBD | TBD | TBD | TBD | TBD | Link | Planned | Yes |
| DAAGA | HSC+ | Optimization-based | TBD | TBD | TBD | TBD | TBD | TBD | Link | Planned | Yes |
| GA | QWS | Optimization-based | TBD | TBD | TBD | TBD | TBD | TBD | Link | Yes | TBD |
| LLM Planner | HSC+ | LLM-based | TBD | TBD | TBD | TBD | TBD | TBD | Link | Planned | Yes |
| Multi-agent pipeline | HSC+ | LLM-based | TBD | TBD | TBD | TBD | TBD | TBD | Link | Planned | Yes |