This project aims to create a system that can solve optimization problems in the cloud. Users should be able to submit an optimization problem to be solved, select one or more solvers to solve it in parallel and get the answer along with the solver which got a solution first.
The website can be found at Handvask.
Starting the backend
The backend is a FastAPI application. To start it, run:
Stay in the handvask directory (the home directory for the repository) and run the following commands:
-
If .env is not created create a file called
~/python_backend/.env
and add the following content:- DB_HOST="127.0.0.1"
- DB_USER="username"
- DB_PASS="password"
- DB_NAME="database"
- HANDVASK_FRONTEND_ORIGIN="*"
- MZN_API_KEY="1234"
- MZN_MN_HOST="http://localhost:8383"
Note that this example configuration assumes that a local mysql database is running on localhost.
- Run the following commands:
sudo apt install mysql-server
pip3 install -r requirements.txt
uvicorn python_backend.app.main:app --reload
-
localhost_backend here you can see the API documentation.
-
To run the backend in a docker container, first move the environment to the local shell using export. Note that this process can be simplified by adding
export
before each line in the .env file, and then executingsource .env
in the shell. Once the environment variables are defined, run the following:
cd python_backend
docker compose build
docker compose up
Remember! If you start the backend as a container, and you're running the database on localhost, the DB_HOST
can't just be localhost, as the localhost will no longer be the same. Instead something like host.docker.internal
might work.
The frontend is a React and Next application. To start it, go to the directory /handvask_frontend
and run:
touch .env
echo "NEXT_PUBLIC_API_URL=http://127.0.0.1:8080/" > .env
npm install
npm next build
npx next start
If it doesn't work, try npx next dev. This will start the frontend on port 3000.
To start the frontend in a container instead of the local machine, use the export
command to store the environment variables in the shell.
First, navigate to the directory /minizinc
and create the .env file in the /app
directory:
touch .env
echo "export API_KEY=\"1234\"" >> .env
echo "export BACKEND_HOST=\"http://host.docker.internal:8080/minizinc\"" >> .env
echo "export SOLVER_IMAGE=\"handvask-minizinc-solver-image:latest\"" >> .env
source .env
Once the environment variables are set up correctly, run:
docker compose build
This will build the two images needed for the minizinc microservice.
After that, it can be deployed with local kubernetes, e.g. minikube which can be found here: minikube.
If minikube is not running, run the following command:
minikube start
After minikube has started, run the following command:
minikube load image handvask-minizinc-image &&
minikube load image handvask-minizinc-solver-image
These two commands make the images available to the minikube cluster. If this is not done the deployment will not work.
To function locally the image location in minizinc/k8s/deployment.yaml
needs to be changed.
By default it will be ${ARTIFACT_LOCATION_MINIZINC}:${COMMIT_SHA}
due to how it works inside the github action workflow, but locally it should be the actual name of the image (handvask-minizinc-solver-image
).
With everything in place, if using minikube open a separate terminal and run:
minikube tunnel
The minikube tunnel
should run in the background at all times, it might be necessary to input your password, so if it doesn't run
try to wait for a "enter your password" prompt.
Then to deploy the microservice run:
kubectl apply -f k8s/
This will spawn the deployment, clusterrole, clusterrolebinding, and service.
The frontend should be available at Handvask, and API documentation at FastAPI. Additionally, when running locally, Minizinc-app will also be available.
The deploy pipeline is constructed using GitHub actions, which is triggered when a change is pushed to the main branch. The pipeline builds the docker images for the backend, frontend and minizinc, after which they are pushed to the appropriate registry. Finally, the pipeline will deploy the images in the cloud.
The pipeline is split into four workflows and three sections.
- Pull-request
Whenever a local branch change is done, the person responsible for the branch creates a pull request to the main branch. The pull request will then be reviewed by the person responsible for the main branch.
The pull request will also trigger the pipeline to run, which will build the docker images where changes were made. Right now there is only test for the backend when the build for the backend is triggered
If only changes were made in the handvask_frontend, the pipeline will catch this in:
- name: Control if changes in handvask_frontend been made
uses: dorny/paths-filter@v2
id: changes
with:
filters: |
backend:
- 'handvask_frontend/**'
This is to save on resources when possible, and produces a faster feedback loop. If the build doesn't fail, the pull request will be reviewed by a member of the group and then merged to the main branch.
- Integration
Whenever a pull-request is approved and merged into the main branch, the integration workflow will run. This workflow will build the docker images for the backend, frontend and minizinc, after which they are pushed to the registry specified by env variables for the respective images, which in this case is the google cloud artifact registry for the project.
Integration has the same constraint as pulls, where the images will only be build if there are changes made since it was built previously.
The integration, in yaml format, can be seen in the file .github/workflows/frontend_CI-CD.yml
The integration part is the second block of the image called frontend image build.
- Deploy
The deploy workflow is triggered whenever the CI part finishes, which then deploys the images to the cloud. The deploy workflow, in yaml format, can be seen in the file .github/workflows/frontend_CI-CD.yml. After a successful rollout, the cypress tests are performed on the production website.
- Google Account
You will need a google account to deploy the pipeline. You can create one here: Google
- Github Account
When you have created your brand new google account. You'll need to create a project in github, for this you'll need to go to the following link: GitHub and create a new repository.
- Fork Handvask the repository
- Clone your forked repository to your local machine.
- Go to the settings of the repository and go to the secrets tab.
- Create the database in cloud console.
- Browse to SQL
- Click "Create instance"
- Select MySQL (currently the backend is hardcoded to expect mysql.)
- Choose some instance name and password for the root user
- Make sure to use the same region as everywhere else (e.g.
europe-north1
) - Once it's up, don't forget to create a database, and user account that can access said database.
- Once you have a VPC created (step 14), don't forget to enable private networking for the database in edit -> connections -> private network, set the private network to the VPC.
NOTE The database does not have to be in gcloud. It can be whereever you want, as long as it's a Mysql Database, and the backend is able to access it using the specified environment variables, it will work.
- Create the following secrets:
{
"total_count": 15,
"secrets": [
{
"name": "API_KEY",
"description": "The API key used for communication between minizinc and the backend. We recommend a string of at least 32 characters, a good mix of lowercase, upper case and numbers should be sufficient."
},
{
"name": "ARTIFACT_LOCATION_BACKEND",
"description": "Artifacts location for Backend. Used for tagging, pushing, pulling and deploying images."
},
{
"name": "ARTIFACT_LOCATION_FRONTEND",
"description": "Artifacts location for Frontend. Used for tagging, pushing, pulling and deploying images."
},
{
"name": "ARTIFACT_LOCATION_MINIZINC",
"description": "Artifacts location for miniznc. Used for tagging, pushing, pulling and deploying images."
},
{
"name": "ARTIFACT_LOCATION_MINIZINC_SOLVER",
"description": "Artifacts location for Minizinc solver image. Used for tagging, pushing, pulling and deploying images."
},
{
"name": "BACKEND_HOST",
"description": "The hostname of the backend, used by the minizinc-app-service. Note that this value should end with /minizinc, e.g. if you're running the backend on localhost and the solver in a local kubernetes cluster, the HOST could be 'http://host.docker.internal:8080/minizinc'."
},
{ "name": "DB_HOST", "description": "The hostname of the database." },
{ "name": "DB_NAME", "description": "The name of the database to use" },
{
"name": "DB_USER",
"description": "THe username of an db user that can access the given database"
},
{ "name": "DB_PASS", "description": "The password for the db user." },
{ "name": "GKE_CLUSTER_NAME", "description": "Cluster name in the GKE" },
{
"name": "GKE_PROJECT",
"description": "The project name in the Google cloud"
},
{ "name": "GKE_REGION", "description": "Region of deployment" },
{
"name": "GOOGLE_CREDENTIALS",
"description": "Googe credentials which will help you run things from Github actions"
},
{
"name": "NEXT_PUBLIC_API_URL",
"description": "The URL of the API, used by the frontend."
}
]
}
-
Create a service account in Google cloud.
-
Workpool. Follow the guide provided on the link Workpool to create a workpool for your service account to use for the github actions workflows
-
Change the workflow authentication so it works with your service accounts
You must now change all the workflow files so they match your service accounts and workflow pool.
This you must do for all three deployment files
backend_CI-CD.yml
frontend_CI-CD.yml
minizinc_CI-cd.yml
- VPC Network in google cloud
For the application to work you'll to create your own VPC network and under the the cloud run you have to create a Serverless VPC access connector, and route only private request through the VPC.
- Create Cloud Run.
Create a new cloud run in google cloud and import the enviroment variable the console
Once it's created, don't forget to go to edit and deploy revision -> connections, and add the cloud run to VPC network, as well as add a SQL connecetion to the datebase instance you set up earlier.
- Create GKE cluster.
Go to the Kubernetes engine in the google cloud and create a cluster.
This is the name for the cluster which should be provided in the github workflow secrets GKE_CLUSTER_NAME
- Integration and deploy
Make a slight change in each of the following directories (e.g. add an empty file):
handvask_frontend/
python_backend/
minizinc/
This will trigger the workflows and deploy the application.
- Fold your hands, and pray to your diety of choice. 🙏
The application is split into three parts, the frontend, the backend, and the minizinc. The frontend is the part that the user interacts with. The backend is the part that handles the communication between the frontend and the minizinc. The minizinc is the part that handles the solving of the optimization problems.
What happens when a user submits a problem to be solved?
- The user submits the problem to the frontend.
- The frontend sends the problem to the backend.
- The backend sends the problem to the minizinc.
- The minizinc starts a job with the specified problem and the solver sends from the backend, which the backend received from the frontend. It will start a job for every solver specified.
- When a solver has found a solution, the minizinc sends the result back to the backend. Whether it is the optimal solution or not.
- The frontend will then request updated information from the backend.