Skip to content

Load test different models in the inference-server #1628

Open
@jackapbutler

Description

@jackapbutler

We want to test the performance of different models within the inference server to understand how it scales with model size such as;

Metadata

Metadata

Assignees

No one assigned

    Labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions