Skip to content

Latest commit

 

History

History
233 lines (146 loc) · 8.8 KB

README.rst

File metadata and controls

233 lines (146 loc) · 8.8 KB

Koku README

license Build Status codecov Updates Python 3 Docs

About

Koku's goal is to provide an open source solution for cost management of cloud and hybrid cloud environments. This solution is offered via a web interface that exposes resource consumption and cost data in easily digestible and filterable views. The project also aims to provide insight into this data and ultimately provide suggested optimizations for reducing cost and eliminating unnecessary resource usage.

Full documentation is available through readthedocs.

Getting Started

This project is developed using Python 3.6. Make sure you have at least this version installed.

Prerequisites

  • Docker
  • PostgreSQL

For Mac OSX

Install Docker for Mac

Install brew

Install PostgreSQL:

brew install postgresql

Development

To get started developing against Koku first clone a local copy of the git repository.

git clone https://github.com/project-koku/koku

This project is developed using the Django web framework. Many configuration settings can be read in from a .env file. To configure, do the following:

  1. Copy example.env into a .env

  2. Obtain AWS values and update the following in your .env:

    AWS_ACCESS_KEY_ID=YOUR_AWS_ACCESS_KEY
    AWS_SECRET_ACCESS_KEY=YOUR_AWS_SECRET_KEY
    AWS_RESOURCE_NAME=YOUR_COST_MANAGEMENT_AWS_ARN
    
  3. (Mac Only) If you are on Mac, do the following note that psycopg2 is a dependency of Django and installing the psycopg2 wheel will likely fail. The following steps should be taken to allow installation to succeed:

    brew install openssl
    brew unlink openssl && brew link openssl --force
    
  4. (Mac Only) Also add the following to your .env`:

    LDFLAGS="-L/usr/local/opt/openssl/lib"
    CPPFLAGS="-I/usr/local/opt/openssl/include"
    
  5. Developing inside a virtual environment is recommended. A Pipfile is provided. Pipenv is recommended for combining virtual environment (virtualenv) and dependency management (pip). To install pipenv, use pip

    pip3 install pipenv
    
  6. Then project dependencies and a virtual environment can be created using

    pipenv install --dev
    
  7. If dependency installation still fails, try using

    pipenv install --dev --sequential
    
  8. To activate the virtual environment run

    pipenv shell
    
  9. Install the pre-commit hooks for the repository

    pre-commit install
    

Developing with Docker Compose

This will explain how to start the server and its dependencies using Docker, create AWS/OCP providers, and view reports. This will not cover all API or scenarios but should give you an end to end flow.

Starting Koku using Docker Compose

Run the following commands:

make docker-up
docker-compose logs -f koku-server koku-worker
pip install koku-nise

Run AWS Scenario

  1. Create AWS Provider:

    make aws-provider aws_name=AWS-PROVIDER-001 bucket=cost-usage-bucket
    
  2. Verify provider exists by visiting http://127.0.0.1:8000/api/cost-management/v1/providers/

  3. Trigger MASU processing by visiting http://127.0.0.1:5000/api/cost-management/v1/download/

  4. Wait for processing to complete

  5. Verify data existing using AWS API endpoints

Run OCP Scenario

  1. Create OCP Provider:

    make ocp-provider-from-yaml cluster_id=my_test_cluster srf_yaml=../nise/example_ocp_static_data.yml ocp_name=my_ocp_name
    
  2. Verify provider exists by visiting http://127.0.0.1:8000/api/cost-management/v1/providers/

  3. Trigger MASU processing by visiting http://127.0.0.1:5000/api/cost-management/v1/download/

  4. Wait for processing to complete

  5. Verify data exists using API endpoints

Stopping Koku using Docker Compose

To bring down all the docker containers, run the following command:

make docker-down

Database

PostgreSQL is used as the database backend for Koku. A docker-compose file is provided for creating a local database container. Assuming the default .env file values are used, to access the database directly using psql run

PGPASSWORD=postgres psql postgres -U postgres -h localhost -p 15432

Note: There is a known limitation with docker-compose and Linux environments with SELinux enabled. You may see the following error during the postgres container deployment:

"mkdir: cannot create directory '/var/lib/pgsql/data/userdata': Permission denied" can be resolved by granting ./pg_data ownership permissions to uid:26 (postgres user in centos/postgresql-96-centos7)

If you see this error, run the following command (assuming you are at the project top level directory):

setfacl -m u:26:-wx ./pg_data

See https://access.redhat.com/containers/?tab=overview#/registry.access.redhat.com/rhel8/postgresql-96

Developing with OpenShift

Our production deployment runs on OpenShift. At times you may need to run on OpenShift if you are working on deployment templates or would like to test in a production like environment. This is a more advanced scenario that many new developers will not need. To learn how to run OpenShift refer to Working with Openshift.

Testing

Koku uses tox to standardize the environment used when running tests. Essentially, tox manages its own virtual environment and a copy of required dependencies to run tests. To ensure a clean tox environment run

tox -r

This will rebuild the tox virtual env and then run all tests.

To run unit tests specifically:

tox -e py36

To run a specific subset of unit tests, you can pass a particular module path to tox. To do this, use positional args using the -- separator. For example:

tox -e py36 -- masu.test.external.downloader.azure.test_azure_services.AzureServiceTest

To run IQE Smoke, Vortex or API tests, while on the Red Hat network and koku deployed via docker-compose run:

make docker-iqe-smokes-tests
make docker-iqe-vortex-tests
make docker-iqe-api-tests

Individual IQE tests can be ran with run_test.sh:

<koku_topdir>/testing/run_test.sh iqe tests plugin hccm -k test_api_cost_model_markup_calculation_ocp

Linting

This repository uses pre-commit to check and enforce code style. It uses Black to reformat the Python code and Flake8 to check it afterwards. Other formats and text files are linted as well.

To run pre-commit checks:

pre-commit run --all-files

pgAdmin

If you want to interact with the Postgres database from a GUI:

  1. Copy the pgadmin_servers.json.example into a pgadmin_servers.json file and if necessary, change any variables to match your database.
  2. docker-compose up causes pgAdmin to run on http://localhost:8432
  3. In the login screen, the default login email is postgres

Side note: The pgadmin_servers.json file uses [pgadmin servers.json syntax](https://www.pgadmin.org/docs/pgadmin4/development/import_export_servers.html#json-format)

Contributing

Please refer to Contributing.