Compare commits
113 Commits
cicd
...
3c5206139b
| Author | SHA1 | Date | |
|---|---|---|---|
| 3c5206139b | |||
| 6c53f39a28 | |||
| 64d8f47174 | |||
| f5f3246e89 | |||
| ad8961ab82 | |||
| c2c8630c82 | |||
| dce9e1fc71 | |||
| 2b84bab1d0 | |||
| 303583daa1 | |||
| d07b2ffbfb | |||
| 335800c44b | |||
| 5766cfde47 | |||
| 2ed3fcc0f9 | |||
| bf72295615 | |||
| 6b860f8ea5 | |||
| 2fa2fa547b | |||
| 3de956872c | |||
| 1c8e3992d6 | |||
| e6e9e141c8 | |||
| f8ece06ee8 | |||
| 149a351bfd | |||
| 0164717b8e | |||
| 104952bfc3 | |||
| f96241c8d1 | |||
| ac6b534f58 | |||
| 06910cd69a | |||
| a48ba520fc | |||
| 9f18aa5890 | |||
| ab3bd84f3b | |||
| f829cd5a4c | |||
| 0f2bf95b71 | |||
| 6a307016ec | |||
| 51017ef8fa | |||
| 05560534bc | |||
| c882173e78 | |||
| 1d94211428 | |||
| 37357080d8 | |||
| 5afa13ac92 | |||
| 416cad1c8f | |||
| b5f83b7163 | |||
| 20cfb5f345 | |||
| 88c96b95f2 | |||
| f6c500b02a | |||
| d702cd8716 | |||
| 329cdd4838 | |||
| 1b70024a29 | |||
| 58206853ee | |||
| 6356398c40 | |||
| 8519922d78 | |||
| 5ac0654fd4 | |||
| 6c07a81b4f | |||
| ba45b4f961 | |||
| 280de82a52 | |||
| 6022e2d879 | |||
| 1996efcc0d | |||
| 80569119cb | |||
| 98e71d4e8a | |||
| fec7191ac2 | |||
| 9b1085f206 | |||
| b35d175a5c | |||
| 7f5fb022ac | |||
| 2d3314ab18 | |||
| 8b489f013d | |||
| 16ce5506d8 | |||
| e440bf8372 | |||
| 607db267e6 | |||
| 352ca64e09 | |||
| f2b735da6e | |||
| 6f7cfb713e | |||
| 103b703ee9 | |||
| daf8b1dce6 | |||
| c088affd74 | |||
| ecc727c991 | |||
| 632569fa5d | |||
| 6c6cbb7396 | |||
| 5e6bfdf77e | |||
| 35e5e18b79 | |||
| c0e8c6bd84 | |||
| 64541b76c5 | |||
| f65b9262cb | |||
| 2765d0548e | |||
| 951f810ce5 | |||
| d2c177d448 | |||
| 299727a7b4 | |||
| b97976b2c5 | |||
| 20241661ff | |||
| ad5c0bea67 | |||
| 80a44277bc | |||
| 5c2b5affc9 | |||
| cd99743d1e | |||
| b39432be1a | |||
| 03f9a33e54 | |||
| 699a9c1e76 | |||
| 4dfd02291e | |||
| e7ca485a88 | |||
| 8319cbfe17 | |||
| 4a023e9f10 | |||
| 4100f96dc6 | |||
| ca24f098e4 | |||
| 80dcd62199 | |||
| 0cfd3da728 | |||
| e141851a87 | |||
| 89ec67999b | |||
| ec38daaedc | |||
| 45c0826a84 | |||
| 45a383cf85 | |||
| 90aff209f9 | |||
| 13528e91e9 | |||
| 04179d633c | |||
| 0a241305d3 | |||
| 31565a0bc4 | |||
| af747417d3 | |||
| c6606c4151 |
@@ -1,36 +0,0 @@
|
||||
name: Gitea Actions Demo
|
||||
run-name: ${{ gitea.actor }} is testing out Gitea Actions 🚀
|
||||
on: [push]
|
||||
|
||||
jobs:
|
||||
Run Tests:
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Check out repository code
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Checkout code
|
||||
uses: actions/checkout@v3
|
||||
|
||||
- name: Set up Python
|
||||
uses: actions/setup-python@v4
|
||||
|
||||
- name: Install dependencies if requirements.txt exists
|
||||
run: |
|
||||
python3 -m venv venv
|
||||
source venv/bin/activate
|
||||
if [ -f requirements.txt ]; then
|
||||
pip install --no-deps -r requirements.txt
|
||||
else
|
||||
echo "No requirements.txt found, skipping installation."
|
||||
fi
|
||||
|
||||
- name: Read .env from server and write to current directory
|
||||
run: |
|
||||
cat /home/jhaeder/.env > .env
|
||||
|
||||
|
||||
- name: Run unit tests
|
||||
run: |
|
||||
source venv/bin/activate
|
||||
python3 manage.py test
|
||||
36
Dockerfile
Normal file
36
Dockerfile
Normal file
@@ -0,0 +1,36 @@
|
||||
# Nutze ein schlankes Python-Image
|
||||
FROM python:3.11-slim-bullseye
|
||||
|
||||
ENV PYTHONUNBUFFERED 1
|
||||
|
||||
WORKDIR /konova
|
||||
|
||||
# Installiere System-Abhängigkeiten
|
||||
RUN apt-get update && apt-get install -y --no-install-recommends \
|
||||
gdal-bin redis-server nginx \
|
||||
&& rm -rf /var/lib/apt/lists/* # Platz sparen
|
||||
|
||||
# Erstelle benötigte Verzeichnisse & setze Berechtigungen
|
||||
RUN mkdir -p /var/log/nginx /var/log/gunicorn /var/lib/nginx /tmp/nginx_client_body \
|
||||
&& touch /var/log/nginx/access.log /var/log/nginx/error.log \
|
||||
&& chown -R root:root /var/log/nginx /var/lib/nginx /tmp/nginx_client_body
|
||||
|
||||
# Kopiere und installiere Python-Abhängigkeiten
|
||||
COPY ./requirements.txt /konova/
|
||||
RUN pip install --upgrade pip && pip install --no-cache-dir -r requirements.txt
|
||||
|
||||
# Entferne Standard-Nginx-Site und ersetze sie durch eigene Config
|
||||
RUN rm -rf /etc/nginx/sites-enabled/default
|
||||
COPY ./nginx.conf /etc/nginx/conf.d
|
||||
|
||||
# Kopiere restliche Projektdateien
|
||||
COPY . /konova/
|
||||
|
||||
# Sammle statische Dateien
|
||||
RUN python manage.py collectstatic --noinput
|
||||
|
||||
# Exponiere Ports
|
||||
#EXPOSE 80 6379 8000
|
||||
|
||||
# Setze Entrypoint
|
||||
ENTRYPOINT ["/konova/docker-entrypoint.sh"]
|
||||
56
README.md
56
README.md
@@ -4,6 +4,7 @@ the database postgresql and the css library bootstrap as well as the icon packag
|
||||
fontawesome for a modern look, following best practices from the industry.
|
||||
|
||||
## Background processes
|
||||
### !!! For non-docker run
|
||||
Konova uses celery for background processing. To start the worker you need to run
|
||||
```shell
|
||||
$ celery -A konova worker -l INFO
|
||||
@@ -18,3 +19,58 @@ Technical documention is provided in the projects git wiki.
|
||||
A user documentation is not available (and not needed, yet).
|
||||
|
||||
|
||||
# Docker
|
||||
To run the docker-compose as expected, you need to take the following steps:
|
||||
|
||||
1. Create a database containing docker, using an appropriate Dockerfile, e.g. the following
|
||||
```
|
||||
version: '3.3'
|
||||
services:
|
||||
postgis:
|
||||
image: postgis/postgis
|
||||
restart: always
|
||||
container_name: postgis-docker
|
||||
ports:
|
||||
- 5433:5432
|
||||
volumes:
|
||||
- db-volume:/var/lib/postgresql/data
|
||||
environment:
|
||||
- POSTGRES_PASSWORD=postgres
|
||||
- POSTGRES_USER=postgres
|
||||
networks:
|
||||
- db-network-bridge
|
||||
|
||||
networks:
|
||||
db-network-bridge:
|
||||
driver: "bridge"
|
||||
|
||||
volumes:
|
||||
db-volume:
|
||||
```
|
||||
This Dockerfile creates a Docker container running postgresql and postgis, creates the default superuser postgres,
|
||||
creates a named volume for persisting the database and creates a new network bridge, which **must be used by any other
|
||||
container, which wants to write/read on this database**.
|
||||
|
||||
2. Make sure the name of the network bridge above matches the network in the konova docker-compose.yml
|
||||
3. Get into the running postgis container (`docker exec -it postgis-docker bash`) and create new databases, users and so on. Make sure the database `konova` exists now!
|
||||
4. Replace all `CHANGE_ME_xy` values inside of konova/docker-compose.yml for your installation. Make sure the `SSO_HOST` holds the proper SSO host, e.g. for the arnova project `arnova.example.org` (Arnova must be installed and the webserver configured as well, of course)
|
||||
5. Take a look on konova/settings.py and konova/sub_settings/django_settings.py. Again: Replace all occurences of `CHANGE_ME` with proper values for your installation.
|
||||
1. Make sure you have the proper host strings added to `ALLOWED_HOSTS` inside of django_settings.py.
|
||||
6. Build and run the docker setup using `docker-compose build` and `docker-compose start` from the main directory of this project (where the docker-compose.yml lives)
|
||||
7. Run migrations! To do so, get into the konova service container (`docker exec -it konova-docker bash`) and run the needed commands (`python manage.py makemigrations LIST_OF_ALL_MIGRATABLE_APPS`, then `python manage.py migrate`)
|
||||
8. Run the setup command `python manage.py setup` and follow the instructions on the CLI
|
||||
9. To enable **SMTP** mail support, make sure your host machine (the one where the docker container run) has the postfix service configured properly. Make sure the `mynetworks` variable is xtended using the docker network bridge ip, created in the postgis container and used by the konova services.
|
||||
1. **Hint**: You can find out this easily by trying to perform a test mail in the running konova web application (which will fail, of course). Then take a look to the latest entries in `/var/log/mail.log` on your host machine. The failed IP will be displayed there.
|
||||
2. **Please note**: This installation guide is based on SMTP using postfix!
|
||||
3. Restart the postfix service on your host machine to reload the new configuration (`service postfix restart`)
|
||||
10. Finally, make sure your host machine webserver passes incoming requests properly to the docker nginx webserver of konova. A proper nginx config for the host machine may look like this:
|
||||
```
|
||||
server {
|
||||
server_name konova.domain.org;
|
||||
|
||||
location / {
|
||||
proxy_pass http://localhost:KONOVA_NGINX_DOCKER_PORT/;
|
||||
proxy_set_header Host $host;
|
||||
}
|
||||
}
|
||||
```
|
||||
@@ -12,7 +12,7 @@ from django.contrib.gis import geos
|
||||
from django.urls import reverse
|
||||
|
||||
from api.tests.v1.share.test_api_sharing import BaseAPIV1TestCase
|
||||
from konova.models import Geometry
|
||||
from konova.sub_settings.lanis_settings import DEFAULT_SRID_RLP
|
||||
|
||||
|
||||
class APIV1UpdateTestCase(BaseAPIV1TestCase):
|
||||
@@ -64,7 +64,7 @@ class APIV1UpdateTestCase(BaseAPIV1TestCase):
|
||||
|
||||
put_props = put_body["properties"]
|
||||
put_geom = geos.fromstr(json.dumps(put_body))
|
||||
put_geom = Geometry.cast_to_rlp_srid(put_geom)
|
||||
put_geom.transform(DEFAULT_SRID_RLP)
|
||||
self.assertEqual(put_geom, self.intervention.geometry.geom)
|
||||
self.assertEqual(put_props["title"], self.intervention.title)
|
||||
self.assertNotEqual(modified_on, self.intervention.modified)
|
||||
@@ -94,7 +94,7 @@ class APIV1UpdateTestCase(BaseAPIV1TestCase):
|
||||
|
||||
put_props = put_body["properties"]
|
||||
put_geom = geos.fromstr(json.dumps(put_body))
|
||||
put_geom = Geometry.cast_to_rlp_srid(put_geom)
|
||||
put_geom.transform(DEFAULT_SRID_RLP)
|
||||
self.assertEqual(put_geom, self.compensation.geometry.geom)
|
||||
self.assertEqual(put_props["title"], self.compensation.title)
|
||||
self.assertNotEqual(modified_on, self.compensation.modified)
|
||||
@@ -124,7 +124,7 @@ class APIV1UpdateTestCase(BaseAPIV1TestCase):
|
||||
|
||||
put_props = put_body["properties"]
|
||||
put_geom = geos.fromstr(json.dumps(put_body))
|
||||
put_geom = Geometry.cast_to_rlp_srid(put_geom)
|
||||
put_geom.transform(DEFAULT_SRID_RLP)
|
||||
self.assertEqual(put_geom, self.eco_account.geometry.geom)
|
||||
self.assertEqual(put_props["title"], self.eco_account.title)
|
||||
self.assertNotEqual(modified_on, self.eco_account.modified)
|
||||
@@ -156,7 +156,7 @@ class APIV1UpdateTestCase(BaseAPIV1TestCase):
|
||||
|
||||
put_props = put_body["properties"]
|
||||
put_geom = geos.fromstr(json.dumps(put_body))
|
||||
put_geom = Geometry.cast_to_rlp_srid(put_geom)
|
||||
put_geom.transform(DEFAULT_SRID_RLP)
|
||||
self.assertEqual(put_geom, self.ema.geometry.geom)
|
||||
self.assertEqual(put_props["title"], self.ema.title)
|
||||
self.assertNotEqual(modified_on, self.ema.modified)
|
||||
|
||||
@@ -13,7 +13,7 @@ from django.contrib.gis.geos import GEOSGeometry
|
||||
from django.core.paginator import Paginator
|
||||
from django.db.models import Q
|
||||
|
||||
from konova.models import Geometry
|
||||
from konova.sub_settings.lanis_settings import DEFAULT_SRID_RLP
|
||||
from konova.utils.message_templates import DATA_UNSHARED
|
||||
|
||||
|
||||
@@ -145,8 +145,8 @@ class AbstractModelAPISerializer:
|
||||
if isinstance(geojson, dict):
|
||||
geojson = json.dumps(geojson)
|
||||
geometry = geos.fromstr(geojson)
|
||||
geometry = Geometry.cast_to_rlp_srid(geometry)
|
||||
geometry = Geometry.cast_to_multipolygon(geometry)
|
||||
if geometry.srid != DEFAULT_SRID_RLP:
|
||||
geometry.transform(DEFAULT_SRID_RLP)
|
||||
return geometry
|
||||
|
||||
def _get_obj_from_db(self, id, user):
|
||||
|
||||
@@ -86,7 +86,7 @@ class EcoAccountWorkflowTestCase(BaseWorkflowTestCase):
|
||||
new_title = self.create_dummy_string()
|
||||
new_identifier = self.create_dummy_string()
|
||||
new_comment = self.create_dummy_string()
|
||||
new_geometry = self.create_dummy_geometry()
|
||||
new_geometry = MultiPolygon(srid=4326) # Create an empty geometry
|
||||
test_conservation_office = self.get_conservation_office_code()
|
||||
test_deductable_surface = self.eco_account.deductable_surface + 100
|
||||
|
||||
@@ -103,7 +103,7 @@ class EcoAccountWorkflowTestCase(BaseWorkflowTestCase):
|
||||
"identifier": new_identifier,
|
||||
"title": new_title,
|
||||
"comment": new_comment,
|
||||
"geom": self.create_geojson(new_geometry),
|
||||
"geom": new_geometry.geojson,
|
||||
"surface": test_deductable_surface,
|
||||
"conservation_office": test_conservation_office.id
|
||||
}
|
||||
|
||||
23
docker-compose.yml
Normal file
23
docker-compose.yml
Normal file
@@ -0,0 +1,23 @@
|
||||
version: '3.3'
|
||||
|
||||
services:
|
||||
konova:
|
||||
external_links:
|
||||
- postgis:db
|
||||
- arnova-nginx-server:arnova
|
||||
build: .
|
||||
image: "ksp/konova:1.8"
|
||||
container_name: "konova-docker"
|
||||
command: ./docker-entrypoint.sh
|
||||
restart: always
|
||||
volumes:
|
||||
- /data/apps/konova/uploaded_files:/konova_uploaded_files
|
||||
ports:
|
||||
- "1337:80"
|
||||
|
||||
# Instead of an own, new network, we need to connect to the existing one, which is provided by the postgis container
|
||||
# NOTE: THIS NETWORK MUST EXIST
|
||||
networks:
|
||||
default:
|
||||
external:
|
||||
name: postgis_nat_it_backend
|
||||
27
docker-entrypoint.sh
Executable file
27
docker-entrypoint.sh
Executable file
@@ -0,0 +1,27 @@
|
||||
#!/bin/bash
|
||||
|
||||
set -e # Beende Skript bei Fehlern
|
||||
set -o pipefail # Fehler in Pipelines nicht ignorieren
|
||||
|
||||
# Starte Redis
|
||||
redis-server --daemonize yes
|
||||
|
||||
# Starte Celery Worker im Hintergrund
|
||||
celery -A konova worker --loglevel=info &
|
||||
|
||||
# Starte Nginx als Hintergrundprozess
|
||||
nginx -g "daemon off;" &
|
||||
|
||||
# Setze Gunicorn Worker-Anzahl (Standard: (2*CPUs)+1)
|
||||
WORKERS=${GUNICORN_WORKERS:-$((2 * $(nproc) + 1))}
|
||||
|
||||
# Stelle sicher, dass Logs existieren
|
||||
mkdir -p /var/log/gunicorn
|
||||
touch /var/log/gunicorn/access.log /var/log/gunicorn/error.log
|
||||
|
||||
# Starte Gunicorn als Hauptprozess
|
||||
exec gunicorn --workers="$WORKERS" konova.wsgi:application \
|
||||
--bind=0.0.0.0:8000 \
|
||||
--access-logfile /var/log/gunicorn/access.log \
|
||||
--error-logfile /var/log/gunicorn/error.log \
|
||||
--access-logformat '%({x-real-ip}i)s via %(h)s %(l)s %(u)s %(t)s "%(r)s" %(s)s %(b)s "%(f)s" "%(a)s"'
|
||||
@@ -84,7 +84,7 @@ class EmaWorkflowTestCase(BaseWorkflowTestCase):
|
||||
new_title = self.create_dummy_string()
|
||||
new_identifier = self.create_dummy_string()
|
||||
new_comment = self.create_dummy_string()
|
||||
new_geometry = self.create_dummy_geometry() # Create an empty geometry
|
||||
new_geometry = MultiPolygon(srid=4326) # Create an empty geometry
|
||||
test_conservation_office = self.get_conservation_office_code()
|
||||
|
||||
check_on_elements = {
|
||||
@@ -99,7 +99,7 @@ class EmaWorkflowTestCase(BaseWorkflowTestCase):
|
||||
"identifier": new_identifier,
|
||||
"title": new_title,
|
||||
"comment": new_comment,
|
||||
"geom": self.create_geojson(new_geometry),
|
||||
"geom": new_geometry.geojson,
|
||||
"conservation_office": test_conservation_office.id
|
||||
}
|
||||
self.client_user.post(url, post_data)
|
||||
|
||||
@@ -124,7 +124,7 @@ class EditInterventionFormTestCase(NewInterventionFormTestCase):
|
||||
self.assertIsNotNone(obj.responsible.handler)
|
||||
self.assertEqual(obj.title, data["title"])
|
||||
self.assertEqual(obj.comment, data["comment"])
|
||||
self.assert_equal_geometries(test_geom, obj.geometry.geom)
|
||||
self.assertTrue(test_geom.equals_exact(obj.geometry.geom, 0.000001))
|
||||
|
||||
self.assertEqual(obj.legal.binding_date, today)
|
||||
self.assertEqual(obj.legal.registration_date, today)
|
||||
|
||||
@@ -72,8 +72,9 @@ class SimpleGeomForm(BaseForm):
|
||||
# will be rendered again on failed submit
|
||||
self.initialize_form_field("geom", self.data["geom"])
|
||||
|
||||
# Initialize features list with empty MultiPolygon, so that an empty input will result in a
|
||||
# proper empty MultiPolygon object
|
||||
# Read geojson into gdal geometry
|
||||
# HINT: This can be simplified if the geojson format holds data in epsg:4326 (GDAL provides direct creation for
|
||||
# this case)
|
||||
features = []
|
||||
features_json = geom.get("features", [])
|
||||
accepted_ogr_types = [
|
||||
@@ -101,25 +102,23 @@ class SimpleGeomForm(BaseForm):
|
||||
return is_valid
|
||||
|
||||
is_valid &= self.__is_area_valid(g)
|
||||
g = Polygon.from_ewkt(g.ewkt)
|
||||
is_valid &= g.valid
|
||||
if not g.valid:
|
||||
self.add_error("geom", g.valid_reason)
|
||||
|
||||
polygon = Polygon.from_ewkt(g.ewkt)
|
||||
is_valid &= polygon.valid
|
||||
if not polygon.valid:
|
||||
self.add_error("geom", polygon.valid_reason)
|
||||
return is_valid
|
||||
|
||||
if isinstance(g, Polygon):
|
||||
features.append(g)
|
||||
elif isinstance(g, MultiPolygon):
|
||||
features.extend(list(g))
|
||||
features.append(polygon)
|
||||
|
||||
# Unionize all geometry features into one new MultiPolygon
|
||||
if features:
|
||||
form_geom = MultiPolygon(*features, srid=DEFAULT_SRID_RLP).unary_union
|
||||
else:
|
||||
form_geom = MultiPolygon(srid=DEFAULT_SRID_RLP)
|
||||
form_geom = MultiPolygon(srid=DEFAULT_SRID_RLP)
|
||||
for feature in features:
|
||||
form_geom = form_geom.union(feature)
|
||||
|
||||
# Make sure to convert into a MultiPolygon. Relevant if a single Polygon is provided.
|
||||
form_geom = Geometry.cast_to_multipolygon(form_geom)
|
||||
if form_geom.geom_type != "MultiPolygon":
|
||||
form_geom = MultiPolygon(form_geom, srid=DEFAULT_SRID_RLP)
|
||||
|
||||
# Write unioned Multipolygon into cleaned data
|
||||
if self.cleaned_data is None:
|
||||
|
||||
@@ -11,7 +11,6 @@ from django.contrib.gis.db.models import MultiPolygonField
|
||||
from django.core.exceptions import ObjectDoesNotExist, MultipleObjectsReturned
|
||||
from django.db import models, transaction
|
||||
from django.utils import timezone
|
||||
from django.contrib.gis.geos import MultiPolygon
|
||||
|
||||
from konova.models import BaseResource, UuidModel
|
||||
from konova.sub_settings.lanis_settings import DEFAULT_SRID_RLP
|
||||
@@ -384,36 +383,6 @@ class Geometry(BaseResource):
|
||||
|
||||
return complexity_factor
|
||||
|
||||
@staticmethod
|
||||
def cast_to_multipolygon(input_geom):
|
||||
""" If input_geom is not a MultiPolygon, cast to MultiPolygon
|
||||
|
||||
Args:
|
||||
input_geom ():
|
||||
|
||||
Returns:
|
||||
output_geom
|
||||
"""
|
||||
output_geom = input_geom
|
||||
if not isinstance(input_geom, MultiPolygon):
|
||||
output_geom = MultiPolygon(input_geom, srid=DEFAULT_SRID_RLP)
|
||||
return output_geom
|
||||
|
||||
@staticmethod
|
||||
def cast_to_rlp_srid(input_geom):
|
||||
""" If input_geom is not of RLP SRID (25832), cast to RLP SRID
|
||||
|
||||
Args:
|
||||
input_geom ():
|
||||
|
||||
Returns:
|
||||
output_geom
|
||||
"""
|
||||
output_geom = input_geom
|
||||
if output_geom.srid != DEFAULT_SRID_RLP:
|
||||
output_geom.transform(DEFAULT_SRID_RLP)
|
||||
return output_geom
|
||||
|
||||
|
||||
class GeometryConflict(UuidModel):
|
||||
"""
|
||||
|
||||
@@ -192,11 +192,10 @@ STATICFILES_DIRS = [
|
||||
]
|
||||
|
||||
# EMAIL (see https://docs.djangoproject.com/en/dev/topics/email/)
|
||||
|
||||
# CHANGE_ME !!! ONLY FOR DEVELOPMENT !!!
|
||||
if DEBUG:
|
||||
# ONLY FOR DEVELOPMENT NEEDED
|
||||
EMAIL_BACKEND = 'django.core.mail.backends.filebased.EmailBackend'
|
||||
EMAIL_FILE_PATH = '/tmp/app-messages' # change this to a proper location
|
||||
EMAIL_FILE_PATH = '/tmp/app-messages'
|
||||
|
||||
DEFAULT_FROM_EMAIL = env("DEFAULT_FROM_EMAIL") # The default email address for the 'from' element
|
||||
SERVER_EMAIL = DEFAULT_FROM_EMAIL # The default email sender address, which is used by Django to send errors via mail
|
||||
|
||||
@@ -469,7 +469,7 @@ class BaseTestCase(TestCase):
|
||||
eco_account.save()
|
||||
return eco_account
|
||||
|
||||
def assert_equal_geometries(self, geom1: MultiPolygon, geom2: MultiPolygon, tolerance=0.001):
|
||||
def assert_equal_geometries(self, geom1: MultiPolygon, geom2: MultiPolygon, tolerance = 0.001):
|
||||
""" Assert for geometries to be equal
|
||||
|
||||
Transforms the geometries to matching srids before checking
|
||||
@@ -491,10 +491,7 @@ class BaseTestCase(TestCase):
|
||||
# transformation from one coordinate system into the other, which is valid
|
||||
geom1.transform(geom2.srid)
|
||||
geom2.transform(geom1.srid)
|
||||
self.assertTrue(
|
||||
geom1.equals_exact(geom2, tolerance=tolerance),
|
||||
msg=f"Difference is {abs(geom1.area - geom2.area)} with {geom1.area} and {geom2.area} in a tolerance of {tolerance}"
|
||||
)
|
||||
self.assertTrue(geom1.equals_exact(geom2, tolerance) or geom2.equals_exact(geom1, tolerance))
|
||||
|
||||
|
||||
class BaseViewTestCase(BaseTestCase):
|
||||
|
||||
25
nginx.conf
Normal file
25
nginx.conf
Normal file
@@ -0,0 +1,25 @@
|
||||
server {
|
||||
listen 80;
|
||||
client_max_body_size 25M;
|
||||
|
||||
location / {
|
||||
proxy_pass http://127.0.0.1:8000;
|
||||
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
|
||||
proxy_set_header X-Real-IP $remote_addr;
|
||||
proxy_set_header Host $host;
|
||||
proxy_redirect off;
|
||||
proxy_cache_bypass $http_upgrade;
|
||||
}
|
||||
|
||||
location /static/ {
|
||||
alias /konova/static/;
|
||||
access_log /var/log/nginx/access.log;
|
||||
autoindex off;
|
||||
types {
|
||||
text/css css;
|
||||
application/javascript js;
|
||||
}
|
||||
}
|
||||
|
||||
error_log /var/log/nginx/error.log;
|
||||
}
|
||||
@@ -48,7 +48,7 @@ pytz==2024.2
|
||||
PyYAML==6.0.2
|
||||
qrcode==7.3.1
|
||||
redis==5.1.0b6
|
||||
requests==2.32.3
|
||||
requests<2.32.0
|
||||
six==1.16.0
|
||||
soupsieve==2.5
|
||||
sqlparse==0.5.1
|
||||
|
||||
@@ -15,7 +15,6 @@ class UserNotificationAdmin(admin.ModelAdmin):
|
||||
class UserAdmin(admin.ModelAdmin):
|
||||
list_display = [
|
||||
"id",
|
||||
"sso_identifier",
|
||||
"username",
|
||||
"first_name",
|
||||
"last_name",
|
||||
|
||||
@@ -1,18 +0,0 @@
|
||||
# Generated by Django 5.1.6 on 2025-09-12 06:10
|
||||
|
||||
from django.db import migrations, models
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('user', '0009_user_oauth_token'),
|
||||
]
|
||||
|
||||
operations = [
|
||||
migrations.AddField(
|
||||
model_name='user',
|
||||
name='sso_identifier',
|
||||
field=models.CharField(blank=True, db_comment='Identifies the account based on an unique identifier from the SSO system', max_length=255, null=True),
|
||||
),
|
||||
]
|
||||
@@ -6,7 +6,6 @@ Created on: 15.11.21
|
||||
|
||||
"""
|
||||
from django.contrib.auth.models import AbstractUser
|
||||
from django.core.exceptions import ObjectDoesNotExist
|
||||
|
||||
from django.db import models
|
||||
|
||||
@@ -33,12 +32,6 @@ class User(AbstractUser):
|
||||
db_comment="OAuth token for the user",
|
||||
related_name="+"
|
||||
)
|
||||
sso_identifier = models.CharField(
|
||||
blank=True,
|
||||
null=True,
|
||||
db_comment="Identifies the account based on an unique identifier from the SSO system",
|
||||
max_length=255,
|
||||
)
|
||||
|
||||
def is_notification_setting_set(self, notification_enum: UserNotificationEnum):
|
||||
return self.notifications.filter(
|
||||
@@ -271,48 +264,4 @@ class User(AbstractUser):
|
||||
self.oauth_token.delete()
|
||||
self.oauth_token = token
|
||||
self.save()
|
||||
return self
|
||||
|
||||
@staticmethod
|
||||
def resolve_user_using_propagation_data(data: dict):
|
||||
""" Fetches user from db by the given data from propagation process
|
||||
|
||||
Args:
|
||||
data (dict): json containing user information from the sso system
|
||||
|
||||
Returns:
|
||||
user (User): The resolved user
|
||||
"""
|
||||
username = data.get("username", None)
|
||||
sso_identifier = data.get("sso_identifier", None)
|
||||
if not username and not sso_identifier:
|
||||
raise AssertionError("No username or sso identifier provided")
|
||||
|
||||
try:
|
||||
user = User.objects.get(username=username)
|
||||
except ObjectDoesNotExist:
|
||||
try:
|
||||
user = User.objects.get(sso_identifier=sso_identifier)
|
||||
except ObjectDoesNotExist:
|
||||
raise ObjectDoesNotExist("No user with this username or sso identifier was found")
|
||||
|
||||
return user
|
||||
|
||||
def update_user_using_propagation_data(self, data: dict):
|
||||
""" Update user data based on propagation data from sso system
|
||||
|
||||
Args:
|
||||
data (dict): json containing user information from the sso system
|
||||
|
||||
Returns:
|
||||
user (User): The updated user
|
||||
"""
|
||||
skipable_attrs = {
|
||||
"is_staff",
|
||||
"is_superuser",
|
||||
}
|
||||
for _attr, _val in data.items():
|
||||
if _attr in skipable_attrs:
|
||||
continue
|
||||
setattr(self, _attr, _val)
|
||||
return self
|
||||
return self
|
||||
@@ -44,8 +44,17 @@ class PropagateUserView(View):
|
||||
|
||||
try:
|
||||
status = "updated"
|
||||
user = User.resolve_user_using_propagation_data(body)
|
||||
user = user.update_user_using_propagation_data(body)
|
||||
user = User.objects.get(username=body.get('username'))
|
||||
# Update user data, excluding some changes
|
||||
skipable_attrs = {
|
||||
"username",
|
||||
"is_staff",
|
||||
"is_superuser",
|
||||
}
|
||||
for _attr, _val in body.items():
|
||||
if _attr in skipable_attrs:
|
||||
continue
|
||||
setattr(user, _attr, _val)
|
||||
except ObjectDoesNotExist:
|
||||
user = User(**body)
|
||||
status = "created"
|
||||
|
||||
Reference in New Issue
Block a user