Skip to content

adding run time info to eval and cleaning up output #422

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 3 commits into from
Apr 25, 2024
Merged

Conversation

HDCharles
Copy link
Contributor

@HDCharles HDCharles commented Apr 23, 2024

Summary:

output now includes info on model run time distribution and a cleaned up result output.

also added info to evaluation.md

Test Plan:

python eval.py --checkpoint-path checkpoints/$MODEL_REPO/model.pth
--dtype bfloat16 --device cuda \

Time to run eval: 53.31s.
Time in model.forward: 20.29s, over 186 model evaluations 
forward run time stats - Median: 0.10s Min: 0.04s Max: 2.18s 
For model checkpoints/meta-llama/Llama-2-7b-hf/model.pth 
wikitext:
 word_perplexity,none: 9.1649
 byte_perplexity,none: 1.5133
 bits_per_byte,none: 0.5977
 alias: wikitext

Reviewers:

Subscribers:

Tasks:

Tags:

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Meta Open Source bot. label Apr 23, 2024
Summary:

output now includes info on model run time distribution and a cleaned up
result output.

Test Plan:

python eval.py --checkpoint-path checkpoints/$MODEL_REPO/model.pth \
    --dtype bfloat16 --device cuda \

Time to run eval: 53.31s.
Time in model.forward: 20.29s, over 186 model evaluations
forward run time stats - Median: 0.10s Min: 0.04s Max: 2.18s
For model checkpoints/meta-llama/Llama-2-7b-hf/model.pth
wikitext:
 word_perplexity,none: 9.1649
 byte_perplexity,none: 1.5133
 bits_per_byte,none: 0.5977
 alias: wikitext

Reviewers:

Subscribers:

Tasks:

Tags:
Summary: see added content

Test Plan: n/a

Reviewers:

Subscribers:

Tasks:

Tags:
Summary: removing install instructions

Test Plan:

Reviewers:

Subscribers:

Tasks:

Tags:
Copy link
Contributor

@mikekgfb mikekgfb left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thank you!

@mikekgfb
Copy link
Contributor

runner-et is a known issue, not related to the present PR.

@dbort @larryliu0820

@mikekgfb mikekgfb merged commit f4f315b into main Apr 25, 2024
@mikekgfb mikekgfb deleted the 098_eval_output branch April 25, 2024 07:43
malfet pushed a commit that referenced this pull request Jul 17, 2024
* adding run time info to eval and cleaning up output

Summary:

output now includes info on model run time distribution and a cleaned up
result output.

Test Plan:

python eval.py --checkpoint-path checkpoints/$MODEL_REPO/model.pth \
    --dtype bfloat16 --device cuda \

Time to run eval: 53.31s.
Time in model.forward: 20.29s, over 186 model evaluations
forward run time stats - Median: 0.10s Min: 0.04s Max: 2.18s
For model checkpoints/meta-llama/Llama-2-7b-hf/model.pth
wikitext:
 word_perplexity,none: 9.1649
 byte_perplexity,none: 1.5133
 bits_per_byte,none: 0.5977
 alias: wikitext

Reviewers:

Subscribers:

Tasks:

Tags:

* Adding evaluation.md content

Summary: see added content

Test Plan: n/a

Reviewers:

Subscribers:

Tasks:

Tags:

* docs update

Summary: removing install instructions

Test Plan:

Reviewers:

Subscribers:

Tasks:

Tags:
malfet pushed a commit that referenced this pull request Jul 17, 2024
* adding run time info to eval and cleaning up output

Summary:

output now includes info on model run time distribution and a cleaned up
result output.

Test Plan:

python eval.py --checkpoint-path checkpoints/$MODEL_REPO/model.pth \
    --dtype bfloat16 --device cuda \

Time to run eval: 53.31s.
Time in model.forward: 20.29s, over 186 model evaluations
forward run time stats - Median: 0.10s Min: 0.04s Max: 2.18s
For model checkpoints/meta-llama/Llama-2-7b-hf/model.pth
wikitext:
 word_perplexity,none: 9.1649
 byte_perplexity,none: 1.5133
 bits_per_byte,none: 0.5977
 alias: wikitext

Reviewers:

Subscribers:

Tasks:

Tags:

* Adding evaluation.md content

Summary: see added content

Test Plan: n/a

Reviewers:

Subscribers:

Tasks:

Tags:

* docs update

Summary: removing install instructions

Test Plan:

Reviewers:

Subscribers:

Tasks:

Tags:
malfet pushed a commit that referenced this pull request Jul 17, 2024
* adding run time info to eval and cleaning up output

Summary:

output now includes info on model run time distribution and a cleaned up
result output.

Test Plan:

python eval.py --checkpoint-path checkpoints/$MODEL_REPO/model.pth \
    --dtype bfloat16 --device cuda \

Time to run eval: 53.31s.
Time in model.forward: 20.29s, over 186 model evaluations
forward run time stats - Median: 0.10s Min: 0.04s Max: 2.18s
For model checkpoints/meta-llama/Llama-2-7b-hf/model.pth
wikitext:
 word_perplexity,none: 9.1649
 byte_perplexity,none: 1.5133
 bits_per_byte,none: 0.5977
 alias: wikitext

Reviewers:

Subscribers:

Tasks:

Tags:

* Adding evaluation.md content

Summary: see added content

Test Plan: n/a

Reviewers:

Subscribers:

Tasks:

Tags:

* docs update

Summary: removing install instructions

Test Plan:

Reviewers:

Subscribers:

Tasks:

Tags:
malfet pushed a commit that referenced this pull request Jul 17, 2024
* adding run time info to eval and cleaning up output

Summary:

output now includes info on model run time distribution and a cleaned up
result output.

Test Plan:

python eval.py --checkpoint-path checkpoints/$MODEL_REPO/model.pth \
    --dtype bfloat16 --device cuda \

Time to run eval: 53.31s.
Time in model.forward: 20.29s, over 186 model evaluations
forward run time stats - Median: 0.10s Min: 0.04s Max: 2.18s
For model checkpoints/meta-llama/Llama-2-7b-hf/model.pth
wikitext:
 word_perplexity,none: 9.1649
 byte_perplexity,none: 1.5133
 bits_per_byte,none: 0.5977
 alias: wikitext

Reviewers:

Subscribers:

Tasks:

Tags:

* Adding evaluation.md content

Summary: see added content

Test Plan: n/a

Reviewers:

Subscribers:

Tasks:

Tags:

* docs update

Summary: removing install instructions

Test Plan:

Reviewers:

Subscribers:

Tasks:

Tags:
malfet pushed a commit that referenced this pull request Jul 17, 2024
* adding run time info to eval and cleaning up output

Summary:

output now includes info on model run time distribution and a cleaned up
result output.

Test Plan:

python eval.py --checkpoint-path checkpoints/$MODEL_REPO/model.pth \
    --dtype bfloat16 --device cuda \

Time to run eval: 53.31s.
Time in model.forward: 20.29s, over 186 model evaluations
forward run time stats - Median: 0.10s Min: 0.04s Max: 2.18s
For model checkpoints/meta-llama/Llama-2-7b-hf/model.pth
wikitext:
 word_perplexity,none: 9.1649
 byte_perplexity,none: 1.5133
 bits_per_byte,none: 0.5977
 alias: wikitext

Reviewers:

Subscribers:

Tasks:

Tags:

* Adding evaluation.md content

Summary: see added content

Test Plan: n/a

Reviewers:

Subscribers:

Tasks:

Tags:

* docs update

Summary: removing install instructions

Test Plan:

Reviewers:

Subscribers:

Tasks:

Tags:
malfet pushed a commit that referenced this pull request Jul 17, 2024
* adding run time info to eval and cleaning up output

Summary:

output now includes info on model run time distribution and a cleaned up
result output.

Test Plan:

python eval.py --checkpoint-path checkpoints/$MODEL_REPO/model.pth \
    --dtype bfloat16 --device cuda \

Time to run eval: 53.31s.
Time in model.forward: 20.29s, over 186 model evaluations
forward run time stats - Median: 0.10s Min: 0.04s Max: 2.18s
For model checkpoints/meta-llama/Llama-2-7b-hf/model.pth
wikitext:
 word_perplexity,none: 9.1649
 byte_perplexity,none: 1.5133
 bits_per_byte,none: 0.5977
 alias: wikitext

Reviewers:

Subscribers:

Tasks:

Tags:

* Adding evaluation.md content

Summary: see added content

Test Plan: n/a

Reviewers:

Subscribers:

Tasks:

Tags:

* docs update

Summary: removing install instructions

Test Plan:

Reviewers:

Subscribers:

Tasks:

Tags:
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Meta Open Source bot.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants