Aspis is a tool for creating measurement instruments for AI risks. It helps you systematically analyze and evaluate AI-powered products by converting high-level risk descriptions into specific, measurable concepts that can be operationalized using LLM-as-a-judge evaluation.
Key Features:
- ⚙️ Systematization: Transforms background concepts (product and risk descriptions) into well-defined, measurable systematized concepts
- 🌐 Interactive UI: Streamlit-based interface that guides you through the systematization process with follow-up questions
- 🔗 REST API: Programmatic access for batch evaluations and integration into existing workflows
- ⚖️ LLM-as-a-Judge: Generates ready-to-use prompt templates for evaluating text against specific risk criteria
Aspis uses a systematization methodology to break down abstract AI risks into concrete, evaluable concepts, enabling systematic risk assessment of AI systems. It is based on the methodology described in the paper "Evaluating Generative AI Systems is a Social Science Measurement Challenge", by Wallach et al.
Aspis is hosted on Hugging Face Spaces under the URL below:
https://huggingface.co/spaces/vector-institute/aspis
The API is also available under Hugging Face Spaces. To see the full documentation on the available endpoints, please visit:
https://vector-institute-aspis.hf.space/api/docs
For more details on how to use the API, please see the Using the API section.
To run both the UI and API services using Docker, make sure you have Docker installed then build the image with the command below:
docker build --no-cache -t aspis:latest .Once the image is built, run it with the command below:
docker run --rm -p 8080:8080 aspis:latestPlease refer to the CONTRIBUTING.md file.
Once the application is started using Docker, the UI will be available under http://localhost:8080/.
Upon access, it will ask you for your AI product description and the risk you want to measure in order to produce LLM prompts that can be used to evaluate the product's outputs against the risk (i.e. measurement instruments).
After filling up all the fields, the app will offer the option to download the results as a
.yaml file so you can load the results later or use them in the API (described below).
The API will be available under http://localhost:8080/api..
The main endpoint is http://localhost:8080/api/evaluate_from_file. It is a POST REST API
endpoint that takes a form data with the following fields:
- An string input text
text_to_evaluate - An
openai_api_keyto access the models - A file upload
systematized_concepts_file, which can be downloaded after answering all the questions from the main app.
To see the full documentation for the available endpoints, you can access
http://localhost:8080/api/docs on your browser.