Skip to content

Use sample latency as the metric for llama3.1_8b_edge SingleStream #2324

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Aug 20, 2025

Conversation

pgmpablo157321
Copy link
Contributor

@hanyunfan @arjunsuresh for visibility

@pgmpablo157321 pgmpablo157321 requested a review from a team as a code owner August 13, 2025 20:35
Copy link
Contributor

MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅

@swasson488
Copy link

@pgmpablo157321 to investigate the checks failing.

@anandhu-eng
Copy link
Contributor

Hi @pgmpablo157321 @swasson488 , I think its safe to ignore the failing action as its not related to submission checker. The implementation is trying to access GPU even if the run is CPU only.

Meanwhile, let me check if there is something broken in MLC Scripts.

Copy link
Contributor

@hanyunfan hanyunfan left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks good to me. I will approve it since we need it to be merged asap, and the current issue doesn't caused by the PR itself. But can someone from the automation team fix the python dependency issue in the checks.

@hanyunfan hanyunfan merged commit e20506b into master Aug 20, 2025
40 of 44 checks passed
@github-actions github-actions bot locked and limited conversation to collaborators Aug 20, 2025
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants