diff --git a/app/__init__.py b/app/__init__.py index 70b4cabfe..0bc90e197 100644 --- a/app/__init__.py +++ b/app/__init__.py @@ -1,7 +1,22 @@ from flask import Flask +from flask_sqlalchemy import SQLAlchemy +from flask_migrate import Migrate +db = SQLAlchemy() +migrate = Migrate() def create_app(test_config=None): app = Flask(__name__) + app.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False + app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql+psycopg2://postgres:postgres@localhost:5432/solar_system_dev' + + db.init_app(app) + migrate.init_app(app, db) + + from app.models.planet import Planet + + from .routes.routes import planets_bp + app.register_blueprint(planets_bp) + return app diff --git a/app/models/__init__.py b/app/models/__init__.py new file mode 100644 index 000000000..e69de29bb diff --git a/app/models/planet.py b/app/models/planet.py new file mode 100644 index 000000000..a76e16946 --- /dev/null +++ b/app/models/planet.py @@ -0,0 +1,7 @@ +from app import db + +class Planet(db.Model): + id = db.Column(db.Integer, primary_key=True, autoincrement=True) + name = db.Column(db.String) + description = db.Column(db.String) + revolution_period = db.Column(db.String) \ No newline at end of file diff --git a/app/routes.py b/app/routes.py deleted file mode 100644 index 8e9dfe684..000000000 --- a/app/routes.py +++ /dev/null @@ -1,2 +0,0 @@ -from flask import Blueprint - diff --git a/app/routes/__init__.py b/app/routes/__init__.py new file mode 100644 index 000000000..e69de29bb diff --git a/app/routes/routes.py b/app/routes/routes.py new file mode 100644 index 000000000..08eff56a1 --- /dev/null +++ b/app/routes/routes.py @@ -0,0 +1,139 @@ +from os import abort +from app import db +from app.models.planet import Planet +from flask import Blueprint, jsonify, abort, make_response, request + +planets_bp = Blueprint("planets_bp", __name__, url_prefix="/planets") + +#Error Handling an invalid planet or non-existing planet +def validate_planet(planet_id): + try: + planet_id = int(planet_id) + except: + abort(make_response({"message":f"planet {planet_id} invalid"}, 400)) + + planet = Planet.query.get(planet_id) + + if not planet: + abort(make_response({"message":f"planet {planet_id} not found"}, 404)) + + return planet + +#Create a new Planet +@planets_bp.route("", methods=["POST"]) +def create_planet(): + request_body = request.get_json() + new_planet = Planet(name=request_body["name"], + description=request_body["description"], + revolution_period=request_body["revolution_period"] + ) + + db.session.add(new_planet) + db.session.commit() + + return make_response(f"Planet {new_planet.name} successfully created", 201) + + +# @planets_bp.route("", methods=["GET"]) +# def read_all_planets(): +# planets_response = [] +# planets = Planet.query.all() +# for planet in planets: +# planets_response.append( +# { +# "id": planet.id, +# "name": planet.name, +# "description": planet.description, +# "revolution_period": planet.revolution_period +# } +# ) +# return jsonify(planets_response) + + +# --------------------- Refactoring read_all query to take query params ---------------------------- +@planets_bp.route("", methods=["GET"]) +def read_all_planets(): + # this code replaces the previous (read_all_planets)query all code + # refactoring to stack queries + + planet_query = Planet.query #use variable to stack on queries + + name_query = request.args.get("name") + if name_query: + planet_query = planet_query.filter(Planet.name.ilike(f"%{name_query}%")) + + description_query = request.args.get("description") + if description_query: + planet_query = planet_query.filter(Planet.description.ilike(f"%{description_query}%")) + + planets = planet_query.all() + + + planets_response = [] + for planet in planets: + planets_response.append({ + "id": planet.id, + "name": planet.name, + "description": planet.description, + "revolution_period": planet.revolution_period + }) + + return jsonify(planets_response) + + +#Get a Single(specific) planet endpoint +@planets_bp.route("/", methods=["GET"]) +def read_one_planet(planet_id): + planet = validate_planet(planet_id) + + return { + "id": planet.id, + "name": planet.name, + "description": planet.description, + "revolution_period": planet.revolution_period + } + +#Updating a Planet (endpoint) +@planets_bp.route("/", methods=["PUT"]) +def update_planet(planet_id): + planet = validate_planet(planet_id) + + request_body = request.get_json() + + planet.name = request_body["name"] + planet.description = request_body["description"] + planet.revolution_period = request_body["revolution_period"] + + db.session.commit() + + return make_response(f"Planet #{planet.id} successfully updated") + +#Deleting a Planet (endpoint) +@planets_bp.route("/", methods=["DELETE"]) +def delete_planet(planet_id): + planet = validate_planet(planet_id) + + db.session.delete(planet) + db.session.commit() + + return make_response(f"planet #{planet.id} successfully deleted") + + +# class Planet: +# def __init__(self, id, name, description, revolution_period): +# self.id = id +# self.name = name +# self.description = description +# self.revolution_period = revolution_period + +# PLANETS = [ +# Planet(1, "Mercury", "terrestrial", "87.97 days"), +# Planet(2, "Venus", "terrestrial", "224.7 days"), +# Planet(3, "Earth", "terrestrial", "365.26 days"), +# Planet(4, "Mars", "terrestrial", "1.88 years"), +# Planet(5, "Jupiter", "gaseous", "11.86 years"), +# Planet(6, "Saturn", "gaseous", "29.46 years"), +# Planet(7, "Uranus", "gaseous", "84.01 years"), +# Planet(8, "Neptune", "gaseous", "164.79 years"), +# Planet(9, "Pluto", "icy, rocky", "248.59 years") +# ] diff --git a/migrations/README b/migrations/README new file mode 100644 index 000000000..98e4f9c44 --- /dev/null +++ b/migrations/README @@ -0,0 +1 @@ +Generic single-database configuration. \ No newline at end of file diff --git a/migrations/alembic.ini b/migrations/alembic.ini new file mode 100644 index 000000000..f8ed4801f --- /dev/null +++ b/migrations/alembic.ini @@ -0,0 +1,45 @@ +# A generic, single database configuration. + +[alembic] +# template used to generate migration files +# file_template = %%(rev)s_%%(slug)s + +# set to 'true' to run the environment during +# the 'revision' command, regardless of autogenerate +# revision_environment = false + + +# Logging configuration +[loggers] +keys = root,sqlalchemy,alembic + +[handlers] +keys = console + +[formatters] +keys = generic + +[logger_root] +level = WARN +handlers = console +qualname = + +[logger_sqlalchemy] +level = WARN +handlers = +qualname = sqlalchemy.engine + +[logger_alembic] +level = INFO +handlers = +qualname = alembic + +[handler_console] +class = StreamHandler +args = (sys.stderr,) +level = NOTSET +formatter = generic + +[formatter_generic] +format = %(levelname)-5.5s [%(name)s] %(message)s +datefmt = %H:%M:%S diff --git a/migrations/env.py b/migrations/env.py new file mode 100644 index 000000000..8b3fb3353 --- /dev/null +++ b/migrations/env.py @@ -0,0 +1,96 @@ +from __future__ import with_statement + +import logging +from logging.config import fileConfig + +from sqlalchemy import engine_from_config +from sqlalchemy import pool +from flask import current_app + +from alembic import context + +# this is the Alembic Config object, which provides +# access to the values within the .ini file in use. +config = context.config + +# Interpret the config file for Python logging. +# This line sets up loggers basically. +fileConfig(config.config_file_name) +logger = logging.getLogger('alembic.env') + +# add your model's MetaData object here +# for 'autogenerate' support +# from myapp import mymodel +# target_metadata = mymodel.Base.metadata +config.set_main_option( + 'sqlalchemy.url', + str(current_app.extensions['migrate'].db.engine.url).replace('%', '%%')) +target_metadata = current_app.extensions['migrate'].db.metadata + +# other values from the config, defined by the needs of env.py, +# can be acquired: +# my_important_option = config.get_main_option("my_important_option") +# ... etc. + + +def run_migrations_offline(): + """Run migrations in 'offline' mode. + + This configures the context with just a URL + and not an Engine, though an Engine is acceptable + here as well. By skipping the Engine creation + we don't even need a DBAPI to be available. + + Calls to context.execute() here emit the given string to the + script output. + + """ + url = config.get_main_option("sqlalchemy.url") + context.configure( + url=url, target_metadata=target_metadata, literal_binds=True + ) + + with context.begin_transaction(): + context.run_migrations() + + +def run_migrations_online(): + """Run migrations in 'online' mode. + + In this scenario we need to create an Engine + and associate a connection with the context. + + """ + + # this callback is used to prevent an auto-migration from being generated + # when there are no changes to the schema + # reference: http://alembic.zzzcomputing.com/en/latest/cookbook.html + def process_revision_directives(context, revision, directives): + if getattr(config.cmd_opts, 'autogenerate', False): + script = directives[0] + if script.upgrade_ops.is_empty(): + directives[:] = [] + logger.info('No changes in schema detected.') + + connectable = engine_from_config( + config.get_section(config.config_ini_section), + prefix='sqlalchemy.', + poolclass=pool.NullPool, + ) + + with connectable.connect() as connection: + context.configure( + connection=connection, + target_metadata=target_metadata, + process_revision_directives=process_revision_directives, + **current_app.extensions['migrate'].configure_args + ) + + with context.begin_transaction(): + context.run_migrations() + + +if context.is_offline_mode(): + run_migrations_offline() +else: + run_migrations_online() diff --git a/migrations/script.py.mako b/migrations/script.py.mako new file mode 100644 index 000000000..2c0156303 --- /dev/null +++ b/migrations/script.py.mako @@ -0,0 +1,24 @@ +"""${message} + +Revision ID: ${up_revision} +Revises: ${down_revision | comma,n} +Create Date: ${create_date} + +""" +from alembic import op +import sqlalchemy as sa +${imports if imports else ""} + +# revision identifiers, used by Alembic. +revision = ${repr(up_revision)} +down_revision = ${repr(down_revision)} +branch_labels = ${repr(branch_labels)} +depends_on = ${repr(depends_on)} + + +def upgrade(): + ${upgrades if upgrades else "pass"} + + +def downgrade(): + ${downgrades if downgrades else "pass"} diff --git a/migrations/versions/303fc3f68f6d_adds_planet_model.py b/migrations/versions/303fc3f68f6d_adds_planet_model.py new file mode 100644 index 000000000..a38c10ff6 --- /dev/null +++ b/migrations/versions/303fc3f68f6d_adds_planet_model.py @@ -0,0 +1,34 @@ +"""adds Planet model + +Revision ID: 303fc3f68f6d +Revises: +Create Date: 2022-11-01 13:25:16.340571 + +""" +from alembic import op +import sqlalchemy as sa + + +# revision identifiers, used by Alembic. +revision = '303fc3f68f6d' +down_revision = None +branch_labels = None +depends_on = None + + +def upgrade(): + # ### commands auto generated by Alembic - please adjust! ### + op.create_table('planet', + sa.Column('id', sa.Integer(), autoincrement=True, nullable=False), + sa.Column('name', sa.String(), nullable=True), + sa.Column('description', sa.String(), nullable=True), + sa.Column('revolution_period', sa.String(), nullable=True), + sa.PrimaryKeyConstraint('id') + ) + # ### end Alembic commands ### + + +def downgrade(): + # ### commands auto generated by Alembic - please adjust! ### + op.drop_table('planet') + # ### end Alembic commands ###