Skip to content

Backport the performance improvement from llama.cpp #709

@kovaacs

Description

@kovaacs

It would be very cool if the performance improvements from ggml-org/llama.cpp#613 could be backported to this repo.

I couldn't find an issue for this, if there is one, I'm happy to close this.

Metadata

Metadata

Assignees

No one assigned

    Labels

    performanceCPU and memory usage - results and comparisons

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions