Skip to content
Snippets Groups Projects

Engine API Server

Overview

The Project serves the Engine API and handles state management of multiple Engine instances.

The Engine API stores and provides following information:

  • Playlogs: A history of all audio titles being played by the Engine. This is used for example to generate detailed reports for regulartory purposes.
  • Track Service: Same as track service, but stripped-down information. Used for implementing a track service feature on the radio's website.
  • Active Source: In redundant deployment scenarios the API stores and shares information on which engine instance is currently active. This could be extended to other audio sources.
  • Health Information: In case of some critical issue affecting the functionality of AURA Engine, the history of health status records of the respective engine is stored.
  • Studio Clock: Information on the current and next show to be used in a Studio Clock application.

You can find details on the available API endpoints here: https://app.swaggerhub.com/apis/AURA-Engine/engine-api/1.0.0

Deployment Modes

AURA Engine allows single and redundant deployments for high availability scenarios.

Engine can be deployed and run manually, using Systemd, Supervisor or Docker.

Single Deployment

This is the most simple case. In that scenario the Engine API is deployed on the same host as the Engine itself.

In your live deployment you might not want to expose the API directly on the web. For security reasons it's highly recommended to guard it using something like NGINX, acting as a reverse proxy.

Redundant Deployment

In this scenario there are two Engine instances involved. Here you will need to deploy one Engine API on the host of each Engine instance. Additionally you'll have to set up a third, so-called "Syncronization Node" of the Engine API. This sync instance of Engine API is in charge of synchronizing playlogs and managing the active engine state.

Managing Active Engine State

In order to avoid duplicate playlog storage, the Synchronization Node requires to know what the currently active Engine is. This can be achieved by some external Status Monitor component which tracks the heartbeat of both engines. In case the Status Monitor identifies one Engine as dysfunctional, it sends a REST request to the Sync Node, informing it about the second, functional Engine instance being activated.

The history of active Engine instances is stored in the database of the Sync Node. It is not only used for playlog syncing, but is also handy as an audit log.

At the moment AURA doesn't provide its own Status Monitor solution. You'll need to integrate your self knitted component which tracks the heartbeat of the engines and posts the active engine state to the Sync Node.

High Availability Playlog Synchronization

Usually when some new audio source starts playing, AURA Engine logs it to its local Engine API instance via some REST call. Now, the local API server stores this information in its local database. Next, it also performs a request to the Synchronization API Serve. The Sync Server checks if this request is coming from the currently active engine instance. If yes, it stores this information in the playlog database.

During the synchronization process between some Engine Node and the Synchronization Node, there might be the case, that the latter is not available e.g. due to network outage, maintenance etc. In this situation the playlog obviously can not be synced. That means the local playlog is marked as "not synced". Whenever the Sync Node is up- and running again, some automated job on the Sync Node is continuously checking for "unsynced" records on remote nodes. If there are such records pending to be synced, this job reads them as batches from that Engine Node. To avoid this sync causing high traffic on any engine instance, these batches are read with some configured delay time (see sync_interval, sync_batch_size, and sync_step_sleep in the Sync Node configuration; all values are in seconds).

In your live deployment you might not want to expose the API directly on the web. For security reasons it's highly recommended to guard it using something like NGINX, acting as a reverse proxy.

Getting started

Requirements

If you are not planning to go with Docker or just want to setup a local development environment, then you'll need:

  • Python 3.7+
  • MariaDB
  • Virtualenv

Installation

Create a virtual environment for your Python dependencies:

virtualenv -p python3.7 python

To activate that environment, run

source python/bin/activate

Install the required dependencies

pip3 install -r requirements.txt
# Additional requirements for the chosen database
pip3 install -r contrib/mariadb-requirements.txt
# Create database and user (change password in script)
sudo mysql -u root -p < contrib/mariadb-database.sql

Configuration

Copy the sample configuration file in ./config/sample/sample-production.engine-api to config and edit the file. First update the main configuration and then configure the type of federation. Depending on how you want to run your Engine API node and where it is deployed, you'll needed to uncomment one of these federation sections:

To avoid any malfunction it is important that any other node-type configurations are commented out.

Engine 1 Node

Use this section if you are running AURA Engine standalone or if this is the first API node in a redundant deployment.

Replace api.sync.local with the actual host name or IP of your sync node.

# NODE 1
host_id=1
sync_host="http://api.sync.local:8008"

Engine 2 Node

In case this is the second API node in a redundant deployment.

Replace api.sync.local with the actual host name or IP of your sync node.

# NODE 2
host_id=2
sync_host="http://api.sync.local:8008"

Synchronization Node

This is the synchronization instance in a redundant setup. This instance combines all valid information coming from Engine API 1 and 2.

Replace engine1.local and engine2.local with the actual details of your main nodes.

# NODE SYNC
host_id=0
main_host_1="http://engine1.local:8008"
main_host_2="http://engine2.local:8008"

# The Engine which is seen as "active" as long no other information is received from the status monitor
default_source=1
# How often the Engine 1 and 2 nodes should be checked for unsynced records (in seconds)
sync_interval=3600
# How many unsynced records should be retrieved at once (in seconds)
sync_batch_size=100
# How long to wait until the next batch is requested (in seconds)
sync_step_sleep=2

Running the Server

Development

To run the API in an local development server execute:

./run.sh api-dev

For convenience running a plain ./run.sh also starts the development server.

When you'll need to run all three nodes to do testing during development you can run:

./run.sh api-test-0 # Sync Node
./run.sh api-test-1 # Node 1
./run.sh api-test-2 # Node 2

Here the run script uses the configurations located in ./test/config.

Production

To run the server for production using Gunicorn, you first need to create the Gunicorn configuration by copying the sample ./config/sample/gunicorn/sample-production.gunicorn.conf.py to your config directory.

Then run this from the root directory:

./run.sh api

If this is succeeding, you can now proceed to configure Engine API to run as a system daemon using Systemd or Supervisor.

Running with Systemd

The Systemd unit file configuration expects to be running under the user engineuser. To create such user type:

    sudo adduser engineuser
    sudo adduser engineuser sudo

Copy the systemd unit file in ./config/sample/systemd to /etc/systemd/system. This configuration file is expecting you to have Engine API installed under /opt/aura/engine-api and engineuser owning the files.

Next login to engineuser and give it permissions to the unit file

  su engineuser
  sudo chmod 644 /etc/systemd/system/aura-engine-api.service

Let's start the service

sudo systemctl start aura-engine-api

And check if it has started successfully

sudo systemctl status aura-engine-api

If you experience issues and need more information, check the syslog while starting the service

tail -f /var/log/syslog

You can stop or restart the service with one of these

sudo systemctl stop aura-engine-api
sudo systemctl restart aura-engine-api

Note, any requirements from the Installation step need to be available for that user.

Running with Supervisor

Alternatively to Systemd you can start Engine API using Supervisor. In ./config/sample/supervisor/aura-engine-api.conf you can find an example Supervisor configuration file. Follow the initial steps of the Systemd setup.

Running with Docker

Having the configuration files engine-api.ini and gunicorn.conf.py located in ./config/docker, you can run the API server in a Docker container this way

exec sudo docker run \
    --network="host" \
    --name engine-api \
    --rm -d \
    -u $UID:$GID \
    -v "$BASE_D":/srv \
    -v "$BASE_D/config/docker":/srv/config \
    --tmpfs /var/log/aura/ autoradio/engine-api

The project also contains a convenience script to get started with a one-liner

# Start up a container
./run.sh docker:api

Development

Using the API

You can find details on the available API endpoints here: https://app.swaggerhub.com/apis/AURA-Engine/engine-api/1.0.0

Adding some entry to the playlog:

curl -d '{ "track_start": "2020-06-27 19:14:00", "track_artist": "Mazzie Star", "track_title": "Fade Into You", "log_source": 1 }' -H "Content-Type: application/json" -X POST http://localhost:8008/api/v1/playlog/store

This newly added entry can be queried using your browser in one of the following ways:

# Get the latest entry
http://localhost:8008/api/v1/trackservice/current
# Get a set of the most recent entries
http://localhost:8008/api/v1/trackservice/
# Filter some specific page (the three most recent entries)
http://localhost:8008/api/v1/trackservice?page=0&limit=3

All other API endpoints are listed in the interactive documentation.

http://localhost:8008/api/v1/ui/

Your OpenAPI definition lives here:

http://localhost:8008/api/v1/openapi.json

Extending the API

The workflow for extending the API follows the API First approach. This means you have to edit the API at https://app.swaggerhub.com/apis/AURA-Engine/engine-api/, then download the python-flask server stubs, and replace & merge the existing generated sources in ./src/rest.

All model files can usually be overwritten. Only controller and test classes need to undergo a merge action.

In the future it might be favorable to use a local Codegen to generate the API artifacts.

Caveat: There is an issue with the generated source related to Python 3.7. Therefore ./src/rest/util.py contains a workaround. Think about that when overwriting the existing file.

Creating a local image

If you are a developer and want to create a local image, run

# Build the image
./run.sh docker:build

Publish new image

If you are developer and want to publish a new image to DockerHub, run

# Releasing the image to DockerHub
./run.sh docker:push

About

Aura Engine API is the API interface for the play-out engine of the Aura Radio Software Suite.

This project is based on a swagger-enabled Flask server using an API First approach. It also uses the Connexion library on top of Flask.

AURA stands for Automated Radio and is a swiss army knife for community radio stations. Beside the Engine it provides Steering (Admin Interface for the radio station), Dashboard (Collaborative scheduling and programme coordination), Tank (Audio uploading, pre-processing and delivery). Read more in the Aura Meta repository or on the specific project pages.

Steering Dashboard Tank Engine
Engine API