LLM-based Fine-grained Conditional Probability Estimation [Huggingface Collection]
Ensure you install all required dependencies and add the current directory to your PYTHONPATH:
conda create -n conditional_prob_llm python=3.12
conda activate conditional_prob_llm
pip install -r requirements.txt
export PYTHONPATH=$(pwd):$PYTHONPATHUse scripts/inference.py for a minimal example of running inference with the model:
python scripts/inference.pyTask implementations for training and evaluation are available in src/tasks/. To run a specific task:
python scripts/run_task.py --config-path <path_to_config>All configurations are stored in configs/ as *.jsonnet files. Some parameters require specification through environment variables.
For training and evaluation tasks (configs in configs/training/ or configs/evaluation/), you can use Hugging Face's accelerate library:
- Set up environment variables with
accelerate config - Run tasks with
accelerate launch scripts/run_task.py --config-path <path_to_config>
To synthesize pseudo labels:
- Use
ReasoningBasedProbExtractorto generate LLM estimations via the vLLM backend - Apply agreement-based filtering using
/scripts/data_synthesis.py