Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

NanoGPT and RTX 4090 #577

Open
ArtHughes opened this issue Dec 2, 2024 · 0 comments
Open

NanoGPT and RTX 4090 #577

ArtHughes opened this issue Dec 2, 2024 · 0 comments

Comments

@ArtHughes
Copy link

Hi All. I'm hoping that someone can explain to me what is going on with nanoGPT and my RTX 4090.
I had a Win11 workstation with a GTX 1080 running an extremely small nanoGPT generating N number of new tokens. I set the seed to a constant value before each generate call. The N tokens were returned in ~7-8 secs every time. This was very consistent and the results were re-producible. All was good. The GTX 1080 ran around 35% loading.
I upgraded to a new Win11 workstation with an RTX 4090 and migrated the same code, same data, and same model. The generate time for the RTX will now vary from 3 secs to 40 secs, and the GPU is running at around 96%.
Can anyone explain to me why this is occurring? The results are good, but the variance in generate time is driving me crazy.
Thanks in advance,
Art

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant