Skip to content

The replication package of the paper "Evaluating the Ability of GPT-4o to Generate Verifiable Specifications in VeriFast" in Forge25

Notifications You must be signed in to change notification settings

gradual-verification/LLM-4-generatingVeriFast-Forge25

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

This folder contains the data and documentation for the evaluation of paper Evaluating the Ability of Large Language Models to Generate Verifiable Specifications in VeriFast, where

  • raw_benchmark/: stores the original benchmark programs of VeriFast, and its readme.md shows how to find our benchmark.
  • IO-pairs/: stores 21 input-output pairs of our benchmarks, and its readme.md shows how to get each output file and 3 input files.
  • scripts/: stores the 2 scripts to do prompt engineering (basic prompting and chain of thought), and its readme.md shows how to run them.
  • raw-pairs/: stores the raw output of GPT-4o for 21 IO-pairs with 2 prompting methods.
  • analysis/: stores the documentation and data of qualitative analysis. Here, qualitative_analysis.md shows the workflow and coding for qualitative analysis; qualitative_analysis.xlsx shows the detailed steps of analysis for each benchmark; result_basic_gpt-4o/ and result_CoT_gpt-4o/ stores the fixed output files for each benchmark.

About

The replication package of the paper "Evaluating the Ability of GPT-4o to Generate Verifiable Specifications in VeriFast" in Forge25

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published