Postgres is an open source relationnal database managment system (RDBMS).
It's totally free and is use in production by many.
In this workshop, you will:
- Create Socrate's relational database schema
- Import test data we've created for you
- Visualize this data using pgAdmin4
- Query some data from the web API of previous workshop
Postgres is your relationnal database where pgAdmin is a simple UI to help you manage your database (query, table creation etc...).
Instead of installing directly Postgres and pgAdmin on your local machine, you will run those in containers.
We've created the docker-compose.yaml for you.
Note: docker-compose is a CLI tool that allows you to run multiple containers, describe in a
docker-compose.yaml
(ordocker-compose.yaml
)file.
Because pgAdmin docker image needs to access postgres docker image, they need to be on the same docker network.
Create a new external docker network for postgres:
# creates the postgres docker network
docker network create socrate-postgres
Then, you can just run containers describe in docker-compose.yaml
by typing:
# start containers describe in the docker-compose.yaml file
# in daemon mode. Meaning, you'll start your containers in background
docker compose up -d
# Creating pgadmin4 ... done
# Creating postgres-14 ... done
Note: two volumes are created to store pgAdmin4 config and all postgres data. those volumes are persisted until you remove them explicitly either:
- with
docker volume rm
command- or with
docker compose down -v
(-v removes also container)
Check if your containers are running:
docker compose ps
# you should see your two containers in the list
You should be able to access PgAdmin on localhost:8040
- usermail: [email protected]
- password: sigl2023
For PostgreSQL instance running locally, credentials are:
- user: sigl2023
- password: sigl2023
Note: default credentials are in the docker-compose file
Once logged in add the local PostgreSQL:
-
Register server: right click on Servers > Register server
-
Enter a connection name (e.g. local)
-
Configure pgAdmin to connect to the postgres container:
- host: postgres-14
- username: sigl2023
- password: sigl2023
- You should see postgres schema and sigl2023 schema in the server dropdown:
Then, let's create Socrate's database core schema, with the following specifications:
- A customer can have one or more addresses
- A producer:
- can deliver many products
- can have one or more address
- A product
- belongs to one or more category
- may or maynot have a discount
Here is the corresponding entity-relation diagram (ERD), without table attributes:
To create database and tables, we've created necessary SQL files, so you just have to run them.
Those SQL scripts lives inside postgresql/scripts/ folder.
This folder exists on your local machine but not inside the postgres container.
Copy scripts to postgres container
# from postgresql/
docker compose cp scripts postgres:/tmp/
- Create SOCRATE database (script uses CREATE DATABASE SQL statement):
# from postgresql/
docker compose exec postgres psql -U sigl2023 -f /tmp/scripts/create-database.sql
- Create SOCRATE's tables (script uses CREATE TABLE SQL statement):
# from postgresql/
docker compose exec postgres psql -U sigl2023 -d socrate -f /tmp/scripts/create-tables.sql
You should see your tables in pgAdmin, under: Server > local > databases > socrate > Schemas > public > Tables
Now we've created all tables, we need to add some data to it.
You will use the COPY command of postgres to import some data that we provided you with.
Note: All test data lives under this
scripts/data
folder. It's a split / generated data based on an from french's gouvernement open dataset. The original dataset for this workshop is "produit sude de france" We generated fake customer and customer addresses by sampling random geo location in a 250km radius from Dijon, France. Those positions are necessary for the computation of the carbon tax (based on the distance between producer address and customer address). We wrote a python3 script to generate those random geo locations Same for product discounts, we just randomly attributed discounts on 20% of the products. We wrote another python3 script to generate fake discounts
Load data from CSV files into socrate
database:
# From postgresql/
# execute the SQL script to import data from CSV files (see. scripts/load-data.sql)
docker compose exec postgres psql -U sigl2023 -d socrate -f /tmp/scripts/load-tables.sql
# This may take a minute...
# Output:
# COPY ...
# COPY ...
# ...
To explore your data, you can directly query some rows using pgAdmin's UI on http://localhost:8040
Important: Make sure to refresh tables from the UI after loading data from previous step:
- go to socrate database > Schemas > public
- right click on Tables > Refresh
If you wish to start from scratch, type the following commands:
docker compose down -v docker network rm socrate-postgresThis will:
- delete both pgadmin and postgres container
- delete assiciated docker volumes
- delete the socrate-postgres network
Objective: Let's create a view to query products on discount.
Note: VIEW and MATERIALIZED VIEWS are differents. MATERIALIZED VIEWS is a PostgreSQL only feature where VIEW are default SQL views. You can read more here: https://www.postgresql.org/docs/14/rules-materializedviews.html
Let's use the SQL query tool from pgadmin4 to query products on discount.
We wish to select the follwing fields:
product.id
: the product ID on discountproduct.name
: the product name on discountproducer.description
: we want to put the producer description for product descriptionimage
: theproduct.url_image
of the productproduct.price
: the original product price without discountdiscount
: the final price corresponding toprice - discount
rounded with 2 decimalsproduct_discount.valid_until
: the deadline date for the discount (discount are only valid until a certain date)
Note: We want to match field names that you have in your backend/src/data/discounts_fr.json This will prevent you from any payload adaption on the frontend side (when rendering products/discounts).
To achieve this selection of table fields, you need to JOIN product
table with product_discount
table on the product id
.
The computation of the final_price
can be done directly in the SELECT
statement using ROUND
and CAST
builtin:
SELECT
...,
ROUND(CAST((product.price - product_discount.discount) AS NUMERIC), 2) as discount
FROM ...
From your query tool in pgadmin4, type the following JOIN query:
SELECT
product.id,
product.name,
product.price,
product.url_image as image,
producer.description,
ROUND(CAST((product.price - product_discount.discount) AS NUMERIC), 2) as discount,
product_discount.valid_until
FROM product
JOIN product_discount
ON product_discount.product_id = product.id
JOIN producer
ON producer.external_id = product.producer_external_id;
That's it! You should obtain all products on discounts.
Now create a new postgresql/scripts/create-views.sql
(same folder as other SQL scripts) file:
CREATE VIEW product_on_discount AS
SELECT
product.id,
product.name,
product.price,
product.url_image as image,
producer.description,
ROUND(CAST((product.price - product_discount.discount) AS NUMERIC), 2) as discount,
product_discount.valid_until
FROM product
JOIN product_discount
ON product_discount.product_id = product.id
JOIN producer
ON producer.external_id = product.producer_external_id;
And run it on your local postgres:
# from postgresql/
docker compose cp scritps postgres:/tmp/
docker compose exec postgres psql -U sigl2023 -f /tmp/scripts/create-views.sql
From your pgAdmin UI, create a new view:
- Right click on
socrate > Schemas > public > Views
menu and selectCreate > View
- Name:
ressource_catalog
- In the
Code
tab, just copy/paste the followingSELECT
statement
Now you can directly query your product_on_discount
view:
- from the query tool, run the following SQL statement:
SELECT * FROM product_on_discount;
Objective:
- create
products_in_category
view with following field selection:
id
: the product unique idname
: the name of the productdescription
: the producer descriptionprice
: the price of the product (ignore discounts)image
: the product's image URLcategoryId
: the ID of the product's categorycategoryName
: the name of the product's category
- add your
CREATE VIEW
statement to thescripts/create-views.sql
script (keep the other view from previous step)
Some hints:
- you'll need 3
JOIN
and start with the product table - use the query tool from pgadmin4 when building your query
- use
product_id
andcategory_id
from product_category table to join products and categories - join
producer.external_id
andproduct.producer_external_id
to get access to the producer's description
Note: creating this view is necessary for next steps of this workshop
Objective: Adapt your groupe's backend to read from PostgreSQL:
- products on discount
- products belonging to a category
- all product categories
From your group's API, you need to install a new node module to interact with Postgres: node-postgres
You will also install dotenv node module to manage database credentials on your different environments (your machine and scaleway).
To install it, like any other node modules, type:
# from your backend/ folder (e.g. for group 13: groupe-13/backend)
# select correct version of node (v16)
nvm use v16
# from your web api repository
npm i --save pg dotenv
From your groupeXX/backend folder, create a new .env
file with:
# inside backend/.env
RDB_HOST=localhost
RDB_PORT=5432
RDB_DATABASE=socrate
RDB_USER=sigl2023
RDB_PASSWORD=sigl2023
Note: RDB stands for Relational DataBase
- Create a new
backend/src/database.js
file with:
const process = require("process");
const { Pool } = require("pg");
const dotenv = require("dotenv");
dotenv.config();
const RDB = {
// Create a pool of connection;
// to control number of concurrent connections.
// We leave default values for now.
// Credentials for the RDB comes from the .env file
pool: new Pool({
host: process.env.RDB_HOST,
port: +process.env.RDB_PORT,
database: process.env.RDB_DATABASE,
user: process.env.RDB_USER,
password: process.env.RDB_PASSWORD,
}),
/**
* Helper to run SQL query on postgres.
* @param {*} sql the query as string
* @returns a list of all rows matching the query
*/
queryMany: async function (sql) {
// Get the next connection available in the pool
const client = await this.pool.connect();
const result = await client.query(sql);
// release the connection
client.release();
return result.rows;
},
/**
* Query a page of products on discount
* that are valid as of today
* @returns a list of discounts
*/
getAvailableDiscounts: async function (page, limit) {
const rows = await this.queryMany(`
SELECT * from product_on_discount
LIMIT ${limit} OFFSET ${page};
`)
return rows;
}
}
module.export {
RDB
};
This new RDB
module contains:
queryMany
: a helper function to help you query many rows providing sql code as stringgetProductDiscounts
which will read a page of product on discount from the view you've created in previous stepdotenv.config()
: will load environment variables from the.env
file created earlier. Environment variables are available in a globalprocess.env
object.
Note: Because we could have many data, your query forces
page
andlimit
argument. This forces API consumer to query a limited amount of data. This is necessary when you want to implement paging in frontend:
- if
page = 0
andlimit = 10
, then the 10 first discounts will be return- if
page = 1
andlimit = 10
, then 11th - 20th disounts will be return See. LIMIT and OFFSET SQL statement
Then adapt your /v1/discounts
route in backend/src/server.js
with:
// ...
const { RDB } = require("./database");
// ...
app.get("/v1/discounts", async (req, res) => {
try {
const discounts = await RDB.getAvailableDiscounts(0, 4);
res.send(discounts);
} catch (e) {
res.send({ error: e.message });
}
});
// ...
async
keyword in the function's definition is necessary since reading data from postgreSQL is asynchronous.await RDB.getAvailableDiscounts(1, 4);
will wait to receive the 4 first available discounts before sending discounts back to the client- if anything goes wrong, server returns the error with the reason.
Then, start your api:
# from backend/
# select correct version of node
nvm use v16
node src/server.js
Start your frontend:
# from frontend/
# select correct version of node
nvm use v16
npm start
Important: Make sure your database is still running!
Login on your frontend, and you should see discounts coming from your postgresql!
Objective: Make your frontend query categories comming from PostgreSQL.
- adapt
backend
:- create a new
getAllCategories()
function to read all data from thecategory
table (no pages / limit necessary) - create it in
backend/src/database.js
's RDB object - adapt your route in
backend/src/server.js
to usegetAllCategories
instead of the mock data inbackend/src/categories_fr.json
- create a new
frontend
should not need any adaptions
Objective: Make your frontend query products in categroy comming from PostgreSQL.
- adapt
backend
:- create a new
getProductsInCategory(categoryId, page, limit)
function to read a page of product matching thecategoryId
from theproducts_in_category
view - create it in
backend/src/database.js
's RDB object - adapt your route in
backend/src/server.js
to usegetProductsInCategory
instead of the mock data inbackend/src/products_fr.json
- create a new
frontend
should not need any adaptions
We deployed a postgres for your application in production.
The production database is not open on internet but only in a private network between your group's scaleway instances.
This means you can access production database only from your group's scaleway VM.
You can connect to your postegres's production database from your local pgadmin4 instance
using an SSH Tunnel
to your scaleway VM.
This means that pgadmin, for every query, will go first to your VM and then from the VM process the SQL queries.
-
Register a new server connection on your pgadmin4 running locally
- Copy your group's id_rsa into your pgadmin4's container:
docker compose cp ~/path/to/ssh_key/group-XX-socra-sigl_id_rsa pgadmin4:/var/lib/pgadmin/storage/socra_sigl.fr/id_rsa
- Make sure pgadmin can read this file:
docker exec --user root pgadmin4 sh -c "chown pgadmin:root /var/lib/pgadmin/storage/socra_sigl.fr/id_rsa"
- Register a new connection from pgAdmin4 (same as on your local postgres)
- in
General
tab:- name: production
- in
Connection
tab:- host: pro.postgres.socra-sigl.fr
- port: 5432
- maintainance database: postgres
- username: socra-group-XX
- password: socra-group-XX
- in the
SSH Tunnel
:- Host: groupXX.socra-sigl.fr
- user: sigl
- select
Identity file
and select the/id_rsa
file
- in
- Copy your group's id_rsa into your pgadmin4's container:
-
You should be able to save and see your production database
-
Create the two views from the previous steps using the pgadmin4 ui:
- go to Servers > production
- go to Databases > socra-group-XX > Schemas > public
- right click on Views > Create view
- in
General
tab- name: product_on_discount
- in the
Code
tab, past only your SELECT statement (without theCREATE VIEW ... AS
)
SELECT product.id, product.name, product.price, product.url_image as image, producer.description, ROUND(CAST((product.price - product_discount.discount) AS NUMERIC), 2) as discount, product_discount.valid_until FROM product JOIN product_discount ON product_discount.product_id = product.id JOIN producer ON producer.external_id = product.producer_external_id;
- in
-
Do the same process to create the
products_in_category
view
- Add the following secrets to your github repository (from github UI):
RDB_HOST
: pro.postgres.socra-sigl.frRDB_PORT
: 5432RDB_DATABASE
: socra-group-XXRDB_USER
: socra-group-XXRDB_PASSWORD
: socra-group-XX
- In your
.github/workflow/deploy.yaml
file, add a new step to create the.env
file withRDB
credentials read from your github account'' s secret variables:
# ...
build-backend:
runs_on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- name: setup .env with github secrets for database
working-directory: backend
run: |
cat <<EOF > .env
RDB_HOST=${{ secrets.RDB_HOST }}
RDB_PORT=${{ secrets.RDB_PORT }}
RDB_DATABASE=${{ secrets.RDB_DATABASE }}
RDB_USER=${{ secrets.RDB_USER }}
RDB_PASSWORD=${{ secrets.RDB_PASSWORD }}
EOF
# ...