Branch: refs/heads/master
Home: https://github.com/xwiki-contrib/ai-llm-benchmark
Commit: d4835b33d6b0203e4a7dd800e7c80ee41c78128e
https://github.com/xwiki-contrib/ai-llm-benchmark/commit/d4835b33d6b0203e4a…
Author: Paul Pantiru <paulp2501(a)gmail.com>
Date: 2024-08-06 (Tue, 06 Aug 2024)
Changed paths:
M Snakefile
M archives/full_nomic_ai_llm_benchmark_archive_20240804_101628.zip
M evaluation_results_graphics/RAG-qa_AnswerRelevancy_bar_chart.png
M evaluation_results_graphics/RAG-qa_AnswerRelevancy_box_plot.png
M evaluation_results_graphics/RAG-qa_ContextualPrecision_bar_chart.png
M evaluation_results_graphics/RAG-qa_ContextualPrecision_box_plot.png
M evaluation_results_graphics/RAG-qa_ContextualRecall_bar_chart.png
M evaluation_results_graphics/RAG-qa_ContextualRecall_box_plot.png
M evaluation_results_graphics/RAG-qa_Faithfulness_bar_chart.png
M evaluation_results_graphics/RAG-qa_Faithfulness_box_plot.png
M evaluation_results_graphics/RAG-qa_grouped_bar_chart.png
M evaluation_results_graphics/average_average_power_draw_grouped_chart.png
M evaluation_results_graphics/average_energy_consumption_grouped_chart.png
M evaluation_results_graphics/average_energy_per_input_token_grouped_chart.png
M evaluation_results_graphics/average_energy_per_output_token_grouped_chart.png
M evaluation_results_graphics/average_energy_per_total_token_grouped_chart.png
M evaluation_results_graphics/summarization_Alignment_bar_chart.png
M evaluation_results_graphics/summarization_Alignment_box_plot.png
M evaluation_results_graphics/summarization_Coverage_bar_chart.png
M evaluation_results_graphics/summarization_Coverage_box_plot.png
M evaluation_results_graphics/summarization_grouped_bar_chart.png
M evaluation_results_graphics/text_generation_grouped_bar_chart.png
M evaluation_results_graphics/text_generation_score_bar_chart.png
M evaluation_results_graphics/text_generation_score_box_plot.png
R reports/evaluation_report_20240804_101339.pdf
A reports/evaluation_report_20240806_161956.pdf
M scripts/results_visualization/generate_plots.py
M scripts/results_visualization/generate_report.py
Log Message:
-----------
LLMAI-87: Execute the benchmark and document results
* update report generation
To unsubscribe from these emails, change your notification settings at https://github.com/xwiki-contrib/ai-llm-benchmark/settings/notifications