Your Web News in One Place

Help Webnuz

Referal links:

Sign up for GreenGeeks web hosting
December 28, 2021 07:30 pm GMT

Add MongoDB and PostgreSQL in Django using Docker

Originally published in The Pylot

In this post, you'll learn how to integrate multiple databases with the Django framework and navigate incoming data using a DB router that automatically writes them to the required database.

Real-world example scenario

Usually, the majority of projects are using relational databases such as Postgres or MySQL but sometimes we also need NoSQL databases to hold extra heavy data which decrease the overload of relational databases.

Assume that your project generates tons of logs while processing some heavy tasks in a queue. These log objects must be stored in non-relational databases instead of hitting relational databases each time and extremely overloading them with huge messy log objects. I guess you spot the problem here, so let's take a look at what can we do about it...

Setting up environment

Create an empty directory named app then let's start by creating a Dockerfile that will copy our working directory and also install required dependencies for Python and Postgres.

app/Dockerfile

FROM python:3.8-slimRUN apt-get update \    && apt-get upgrade -y \    && apt-get install -y \    build-essential \    libssl-dev \    libffi-dev \    python3-dev \    build-essential \    libjpeg-dev \    zlib1g-dev \    gcc \    libc-dev \    bash \    git \    && pip3 install --upgrade pipENV LIBRARY_PATH=/lib:/usr/libENV PYTHONUNBUFFERED 1ENV PYTHONDONTWRITEBYTECODE 1WORKDIR /appCOPY . /appRUN pip3 --no-cache-dir install -r requirements.txt

These wheel packages will be used while installing and setting up Postgres and other services in the system. Now, we need to add requirements.txt to install the required packages for our project.

app/requirements.txt

celery==5.0.2Django==3.1.3git+git://github.com/thepylot/djongo.git#egg=djongomongoengine==0.20.0pylint-mongoengine==0.4.0pymongo==3.11.0psycopg2-binary==2.8.5redis==3.5.3Faker

We are going to use djongo which will help to convert SQL to MongoDB query. By using djongo we can use MongoDB as a backend database for our Django project. At the time of writing this post, djongo has issues supporting Django versions above v.3+, but it can be resolved easily by just by changing the version from the package itself by forking it to your repository. There is a file named setup.py that holds configuration settings and you'll see the block named install_requires where the supported version is mentioned (Don't forget to fork it first).

setup.py

install_requires = [    'sqlparse==0.2.4',    'pymongo>=3.2.0',    'django>=2.1,<=3.0.5',]

We just need to refactor it to fit with our current version of Django. I am using 3.1.3 so I will replace the 3.0.5 to 3.1.3 and it will look like below:

install_requires = [    'sqlparse==0.2.4',    'pymongo>=3.2.0',    'django>=2.1,<=3.1.3',]

Once you finished, search for requirements.txt and change the Django version there as well:

requirements.txt (in djongo)

...django>=2.0,<=3.1.3...

That's it! Commit your changes to the forked repository and then you need to include it to your requirements.txt file but this time it will get from your forked repository like below:

git+git://github.com/YOUR_GITHUB_USERNAME/djongo.git#egg=djongo

You don't have to go through all these processes because I already included it as you can see from above (requirements.txt) so feel free to use mine if your Django version is 3.1.3 as well.

There are some other dependencies like celery where we'll use it as the queue to pass time-consuming tasks to run in the background and redis is just a message broker for celery. This topic is out of scope for this post but you can visit Dockerizing Django with Postgres, Redis and Celery to understand it more.

Now it's time to set up our services by configuring compose file. Create this file in a root level of your current directory which means it will be outside of app directory There are going to be 5 services in total:

  • mongodb - for setting up MongoDB
  • postgres - for setting up PostgreSQL
  • app - Django project
  • celery - Queue for tasks
  • redis - Message broker that required for celery

/docker-compose.yml

version: '3'services:  mongo:    image: mongo    container_name: mongo    restart: always    env_file: .env    environment:       - MONGO_INITDB_ROOT_USERNAME=root      - MONGO_INITDB_ROOT_PASSWORD=root      - MONGO_INITDB_DATABASE=${MONGO_DB_NAME}      - MONGO_INITDB_USERNAME=${MONGO_DB_USERNAME}      - MONGO_INITDB_PASSWORD=${MONGO_DB_PASSWORD}    volumes:      - ${PWD}/_data/mongo:/data/db      - ${PWD}/docker/_mongo/fixtures:/import      - ${PWD}/docker/_mongo/scripts/init.sh:/docker-entrypoint-initdb.d/setup.sh    ports:      - 27017:27017  postgres:    container_name: postgres    image: postgres:12    restart: always    env_file: .env    environment:      - POSTGRES_DB=app_db      - POSTGRES_USER=app_db_user      - POSTGRES_PASSWORD=supersecretpassword      - POSTGRES_PORT=5432    ports:      - 5432:5432    volumes:      - ${PWD}/_data/postgres:/var/lib/postgresql/data      - ${PWD}/docker/_postgres/scripts/create_test_db.sql:/docker-entrypoint-initdb.d/docker_postgres_init.sql  redis:    image: redis:6    container_name: redis    restart: always    env_file: .env    command: redis-server --requirepass $REDIS_PASSWORD    ports:      - 6379:6379    volumes:      - ${PWD}/_data/redis:/var/lib/redis  app:    build: ./app    image: app:latest    container_name: app    restart: always    command: "python manage.py runserver 0.0.0.0:8000"    env_file: .env    volumes:      - ${PWD}/app:/app    ports:      - 8000:8000    depends_on:      - postgres      - redis  celery:    build: ./app    image: app:latest    container_name: celery    restart: always    command: [      "celery",      "-A",      "app",      "worker",      "-c",      "1",      "-l",      "INFO",      "--without-heartbeat",      "--without-gossip",      "--without-mingle",    ]    env_file: .env    environment:      - DJANGO_SETTINGS_MODULE=app.settings      - DJANGO_WSGI=app.wsgi      - DEBUG=False    volumes:      - ${PWD}/app:/app    depends_on:      - postgres      - redisnetworks:  default:

I will not go through this configuration in detail by assuming you already have knowledge about docker and compose files. Simply, we are pulling required images for the services and setting up main environment variables and ports to complete the configuration.

Now we also need to add .env file to fetch values of environment variables while building services:

/.env

# Mongo DBMONGO_DB_HOST=mongoMONGO_DB_PORT=27017MONGO_DB_NAME=mongo_dbMONGO_DB_USERNAME=rootMONGO_DB_PASSWORD=rootMONGO_DB_URI=mongodb://root:root@mongo:27017# PostgreSQLPOSTGRES_HOST=postgresPOSTGRES_DB=app_dbPOSTGRES_USER=app_db_userPOSTGRES_PASSWORD=supersecretpasswordPOSTGRES_PORT=5432# RedisREDIS_HOST=redisREDIS_PORT=6379REDIS_PASSWORD=supersecretpasswordBROKER_URL=redis://:supersecretpassword@redis:6379/0REDIS_CHANNEL_URL=redis://:supersecretpassword@redis:6379/1CELERY_URL=redis://:supersecretpassword@redis:6379/0

Next, we'll create a new Django project inside our app folder:

docker-compose run app sh -c "django admin startproject app ."

The project structure should be like below:

. app    app       asgi.py       __init__.py       settings.py       urls.py       wsgi.py    Dockerfile    manage.py    requirements.txt docker-compose.yml .env

If you see dbsqlite in project files then you should delete it since we'll use postgres it as a relational database. You also will notice _data directory which represents the volume of MongoDB and Postgres.

Integration with PostgreSQL

We are ready to add our primary relational database which is going to be postgres. Navigate to settings.py and update DATABASES configuration like below:

DATABASES = {    'default': {        'ENGINE': 'django.db.backends.postgresql',        'HOST': os.environ.get('POSTGRES_HOST'),        'NAME': os.environ.get('POSTGRES_NAME'),        'USER': os.environ.get('POSTGRES_USER'),        'PASSWORD': os.environ.get('POSTGRES_PASSWORD'),        'PORT': os.environ.get('POSTGRES_PORT', 5432),    }}

The default database set to postgres and environment variables will be fetched from .env file. Sometimes, postgres is having connection issues caused by racing issues between Django and postgres. To prevent such situations we'll implement a custom command and add it to commands block in compose file. In this way, Django will wait postgres before launch.

The recommended path of holding commands is /core/management/commands/ from the official documentation of Django. So let's create an app named core then create a management/commands directory inside it.

docker-compose run sh -c "django-admin startapp core"

Then add following command to hold Django until postgres is available:

/core/management/commands/wait_for_db.py

import timefrom django.db import connectionsfrom django.db.utils import OperationalErrorfrom django.core.management import BaseCommandclass Command(BaseCommand):    """Django command to pause execution until db is available"""    def handle(self, *args, **options):        self.stdout.write('Waiting for database...')        db_conn = None        while not db_conn:            try:                db_conn = connections['default']            except OperationalError:                self.stdout.write('Database unavailable, waititng 1 second...')                time.sleep(1)        self.stdout.write(self.style.SUCCESS('Database available!'))

Make sure you included __init__.py into sub-directories you created. Now update app service in compose file by adding this command:

docker-compose.yml

  app:    build: ./app    image: app:latest    container_name: app    restart: always    command: >        sh -c "python manage.py wait_for_db &&               python manage.py migrate &&               python manage.py runserver 0.0.0.0:8000"    env_file: .env    volumes:      - ${PWD}/app:/app    ports:      - 8000:8000    depends_on:      - postgres      - redis

Consider the command block only and you'll see we now have two more commands there.

Integration with MongoDB

Actually, the integration of MongoDB is so simple thanks to djongo which handles everything behind the scenes. Switch to settings.py again and we'll add our second database as nonrel which stands for the non-relational database.

settings.py

DATABASES = {    'default': {        'ENGINE': 'django.db.backends.postgresql',        'HOST': os.environ.get('POSTGRES_HOST'),        'NAME': os.environ.get('POSTGRES_NAME'),        'USER': os.environ.get('POSTGRES_USER'),        'PASSWORD': os.environ.get('POSTGRES_PASSWORD'),        'PORT': os.environ.get('POSTGRES_PORT'),    },    "nonrel": {        "ENGINE": "djongo",        "NAME": os.environ.get('MONGO_DB_NAME'),        "CLIENT": {            "host": os.environ.get('MONGO_DB_HOST'),            "port": int(os.environ.get('MONGO_DB_PORT')),            "username": os.environ.get('MONGO_DB_USERNAME'),            "password": os.environ.get('MONGO_DB_PASSWORD'),        },        'TEST': {            'MIRROR': 'default',        },    }}

The same logic applies here as we did for default database which is postgres.

Setting up DB router

DB router which will automatically write objects to a proper database such as whenever the log object created it should navigate to mongodb instead of postgres. Setting up a DB router is simple, we just need to use router methods that Django provides and define our non-rel models to return the proper database.

Create a new directory named utils inside the core app and also add __init__.py to mark it as a python package. Then add the new file which is DB router below:

/core/utils/db_routers.py

class NonRelRouter:    """    A router to control if database should use    primary database or non-relational one.    """    nonrel_models = {'log'}    def db_for_read(self, model, **_hints):        if model._meta.model_name in self.nonrel_models:            return 'nonrel'        return 'default'    def db_for_write(self, model, **_hints):        if model._meta.model_name in self.nonrel_models:            return 'nonrel'        return 'default'    def allow_migrate(self, _db, _app_label, model_name=None, **_hints):        if _db == 'nonrel' or model_name in self.nonrel_models:            return False        return True

nonrel_models - We are defining the name of our models in lowercase which belongs to a non-rel database or mongodb.

db_for_read - the function name is self-explanatory so basically, it is used for reading operations which means each time we try to get records from the database it will check where it belongs and return the proper database.

db_for_write - the same logic applies here. It's used to pick a proper database for writing objects.

allow_migrate - Decided if the model needs migration. In mongodb there is no need to run migrations since it's a non-rel database.

Next, we should add extra configuration in settings.py to activate our custom router:

settings.py

...DATABASE_ROUTERS = ['core.utils.db_routers.NonRelRouter', ]...

Great! Our database configurations are finished and now it's time to make a few changes in Django as well before launching everything.

Setting up Celery and Redis

This part is a bit out of scope but since I want to illustrate some real-world app then those tools are always present in projects to handle heavy and time-consuming tasks. Let's add celery to our project, but it should place in our project folder alongside with settings file:

celery.py

import osfrom celery import Celeryos.environ.setdefault('DJANGO_SETTINGS_MODULE', 'app.settings')app = Celery('app')app.config_from_object('django.conf:settings', namespace='CELERY')app.autodiscover_tasks()

Basically, it will discover all tasks alongside the project and will pass them to the queue. Next, we also need to update __init__.py file inside the current directory, which is our Django project:

init.py

from .celery import app as celery_app__all__ = ['celery_app']

Celery requires a broker URL for tasks so in this case, we will use Redis as a message broker. Open your settings file and add the following configurations:

settings.py

CELERY_TASK_TRACK_STARTED = TrueCELERY_TASK_TIME_LIMIT = 30 * 60CELERY_IGNORE_RESULT = TrueCELERY_BROKER_URL = os.environ.get('CELERY_URL')CELERYD_HIJACK_ROOT_LOGGER = FalseREDIS_CHANNEL_URL = os.environ.get('REDIS_CHANNEL_URL')

Now try to run docker-compose up -d and all services should start successfully.

Adding models and celery tasks

In this section, we'll pass a new task to the celery queue and test if our DB router works properly and writes objects to the required database. Create a new directory inside the core app named models and add the following file inside it to hold MongoDB models:

core/models/mongo_models.py

from djongo import models as mongo_modelsclass Log(mongo_models.Model):    _id = mongo_models.ObjectIdField()    message = mongo_models.TextField(max_length=1000)    created_at = mongo_models.DateTimeField(auto_now_add=True)    updated_at = mongo_models.DateTimeField(auto_now=True)    class Meta:        _use_db = 'nonrel'        ordering = ("-created_at", )    def __str__(self):        return self.message

As you see it's a very simple Log model where it will let us know what's going on behind the scenes of internal operations of our application. Then add a model for Postgres as well:

core/models/postgres_models.py

from django.db import modelsclass Post(models.Model):    title = models.CharField(max_length=255)    description = models.TextField()    def __str__(self):        return self.title

We are going to generate random posts and make some of them fail in order to write error logs for failed data.

Lastly, don't forget to include __init__.py inside models directory:

models/init.py

from .postgres_models import *from .mongo_models import *

Now we need to create a celery task that will generate tons of posts with random values by using Faker generators:

core/tasks.py

import loggingimport randomfrom faker import Fakerfrom .models import Postfrom celery import shared_taskfrom core.utils.log_handlers import LoggingHandlerlogger = logging.getLogger(__name__)logger.addHandler(LoggingHandler())@shared_taskdef create_random_posts():    fake = Faker()    number_of_posts = random.randint(5, 100)    for i in range(number_of_posts):        try:            if i % 5 == 0:                title = None            else:                 title = fake.sentence()            description = fake.text()            Post.objects.create(                title = title,                description = description,            )        except Exception as exc:            logger.error("The post number %s failed due to the %s", i, exc)

By adding if statement there we are forcing some of the posts to fail in order to catch exceptions and write them to mongodb. As you noticed, we are using a custom log handler that will write the log data right away after its produced. So, create a new file named log_handlers.py inside utils folder:

core/utils/log_handlers.py

import loggingfrom core.models import Logclass LoggingHandler(logging.Handler):    """Save log messages to MongoDB    """    def emit(self, record):        Log.objects.create(            message=self.format(record),        )

Great! We are almost ready to launch.

Lastly, let's finish creating a very simple view and URL path to trigger the task from the browser and return JSON a response if the operation succeeded.

core/views.py

from django.http.response import JsonResponsefrom .tasks import create_random_postsdef post_generator(request):    create_random_posts.delay()    return JsonResponse({"success": True})

urls.py

from django.contrib import adminfrom django.urls import pathfrom core.views import post_generatorurlpatterns = [    path('admin/', admin.site.urls),    path('', post_generator)]

Include the core app inside INSTALLED_APPS configuration in settings.py then run the migrations and we're done!

docker-compose run app sh -c "python manage.py makemigrations core"

Now, if you navigate to 127.0.0.1:8000 the task will be passed to the queue and post objects will start to generate.

Try to visit admin and you'll see log objects created successfully and by this way we're avoiding overload postgres while it handles only relational data.

Source Code:

GitHub logo thepylot / django-mongodb-postgres

Integration of multiple databases with Django framework and navigate incoming data using DB router which automatically writes them to required database.

MongoDB and Postgres integration with Django

This project includes multiple database configurations to launch Django project with Postgres and MongoDB.

Getting Started

This project works on Python 3+ and Django 3.1.3. If you want to change the Django version please follow the link below to get information about it.

Add MongoDB and PostgreSQL in Django using Docker

Run the following command to makemigrations:

docker-compose run app sh -c "python manage.py makemigrations core"

then start the containers:

docker-compose up -d

After containers launched navigate to / home path and celery will automatically receive the task.

Support

If you feel like you unlocked new skills, please share them with your friends and subscribe to the youtube channel to not miss any valuable information.


Original Link: https://dev.to/thepylot/add-mongodb-and-postgresql-in-django-using-docker-55j6

Share this article:    Share on Facebook
View Full Article

Dev To

An online community for sharing and discovering great ideas, having debates, and making friends

More About this Source Visit Dev To