From be45356a4a5ede0f585ea0da15d7c9318f1adfb4 Mon Sep 17 00:00:00 2001 From: Cipher Vance Date: Thu, 18 Sep 2025 20:08:24 -0500 Subject: [PATCH] removed diff txt whoopsie --- diff.txt | 915 ------------------------------------------------------- 1 file changed, 915 deletions(-) delete mode 100644 diff.txt diff --git a/diff.txt b/diff.txt deleted file mode 100644 index 7394524..0000000 --- a/diff.txt +++ /dev/null @@ -1,915 +0,0 @@ -diff --git a/.dockerignore b/.dockerignore -index d9b625e..3f2b844 100644 ---- a/.dockerignore -+++ b/.dockerignore -@@ -1,9 +1,13 @@ -+.env - .git --__pycache__/ --*.py[cod] -+.gitignore -+README.md -+Dockerfile -+.dockerignore -+.air.toml -+tmp/ -+.vscode/ -+.idea/ - *.log --!.env --venv/ --.venv/ --dist/ --build/ -\ No newline at end of file -+coverage.out -+rideaware-api -\ No newline at end of file -diff --git a/Dockerfile b/Dockerfile -index c72dff0..7aa361b 100644 ---- a/Dockerfile -+++ b/Dockerfile -@@ -1,53 +1,52 @@ --FROM python:3.10-slim AS builder -- --ENV PYTHONDONTWRITEBYTECODE=1 \ -- PYTHONUNBUFFERED=1 \ -- PIP_NO_CACHE_DIR=1 -+# Build stage -+FROM golang:1.21-alpine AS builder - -+# Set working directory - WORKDIR /app - --RUN apt-get update && apt-get install -y --no-install-recommends \ -- build-essential gcc \ -- && rm -rf /var/lib/apt/lists/* -+# Install git (needed for some Go modules) -+RUN apk add --no-cache git - --COPY requirements.txt . -+# Copy go mod files -+COPY go.mod go.sum ./ - --RUN python -m pip install --upgrade pip && \ -- pip wheel --no-deps -r requirements.txt -w /wheels && \ -- pip wheel --no-deps gunicorn -w /wheels -+# Download dependencies -+RUN go mod download - --FROM python:3.10-slim AS runtime -+# Copy source code -+COPY . . - --ENV PYTHONDONTWRITEBYTECODE=1 \ -- PYTHONUNBUFFERED=1 \ -- PIP_NO_CACHE_DIR=1 \ -- PORT=5000 \ -- WSGI_MODULE=server:app \ -- GUNICORN_WORKERS=2 \ -- GUNICORN_THREADS=4 \ -- GUNICORN_TIMEOUT=60 \ -- FLASK_APP=server.py -+# Build the application -+RUN CGO_ENABLED=0 GOOS=linux go build -a -installsuffix cgo -o rideaware-api . - --WORKDIR /app -+# Production stage -+FROM alpine:latest - --RUN groupadd -g 10001 app && useradd -m -u 10001 -g app app -+# Install ca-certificates for HTTPS requests and timezone data -+RUN apk --no-cache add ca-certificates tzdata - --COPY --from=builder /wheels /wheels --RUN pip install --no-cache-dir /wheels/* && rm -rf /wheels -+# Create non-root user -+RUN addgroup -g 1001 -S appgroup && \ -+ adduser -u 1001 -S appuser -G appgroup - --# Install python-dotenv if not already in requirements.txt --RUN pip install python-dotenv -+# Set working directory -+WORKDIR /home/appuser - --USER app -+# Copy binary from builder stage -+COPY --from=builder /app/rideaware-api . - --COPY --chown=app:app . . -+# Change ownership to non-root user -+RUN chown -R appuser:appgroup /home/appuser - --# Copy .env file specifically --COPY --chown=app:app .env .env -+# Switch to non-root user -+USER appuser - --EXPOSE 5000 -+# Expose port -+EXPOSE 8080 - --HEALTHCHECK --interval=30s --timeout=3s --start-period=10s --retries=3 \ -- CMD python -c "import os,socket; s=socket.socket(); s.settimeout(2); s.connect(('127.0.0.1', int(os.getenv('PORT', '5000')))); s.close()" -+# Health check -+HEALTHCHECK --interval=30s --timeout=3s --start-period=5s --retries=3 \ -+ CMD wget --no-verbose --tries=1 --spider http://localhost:8080/health || exit 1 - --CMD ["sh", "-c", "exec gunicorn $WSGI_MODULE --bind=0.0.0.0:$PORT --workers=$GUNICORN_WORKERS --threads=$GUNICORN_THREADS --timeout=$GUNICORN_TIMEOUT --access-logfile=- --error-logfile=- --keep-alive=5"] -\ No newline at end of file -+# Run the application -+CMD ["./rideaware-api"] -\ No newline at end of file -diff --git a/README.md b/README.md -index cda9868..65ce1b9 100644 ---- a/README.md -+++ b/README.md -@@ -15,53 +15,78 @@ Whether you're building a structured training plan, analyzing ride data, or comp - Ensure you have the following installed on your system: - - - Docker --- Python 3.10 or later --- pip -+- Go 1.21 or later -+- PostgreSQL (for local development, optional) - - ### Setting Up the Project - - 1. **Clone the Repository** - - ```bash -- git clone https://github.com/VeloInnovate/rideaware-api.git -+ git clone https://github.com/rideaware/rideaware-api.git - cd rideaware-api - ``` - --2. **Create a Virtual Environment** -- It is recommended to use a Python virtual environment to isolate dependencies. -+2. **Install Go Dependencies** - - ```bash -- python3 -m venv .venv -+ go mod tidy - ``` - --3. **Activate the Virtual Environment** -- - On Linux/Mac: -- ```bash -- source .venv/bin/activate -- ``` -- - On Windows: -- ```cmd -- .venv\Scripts\activate -- ``` -- --4. **Install Requirements** -- Install the required Python packages using pip: -+3. **Build the Application** - ```bash -- pip install -r requirements.txt -+ go build -o rideaware-api - ``` - - ### Configuration - - The application uses environment variables for configuration. Create a `.env` file in the root directory and define the following variables: - -+```env -+# Database Configuration -+PG_HOST=your_postgres_host -+PG_PORT=5432 -+PG_DATABASE=rideaware -+PG_USER=your_postgres_user -+PG_PASSWORD=your_postgres_password -+ -+# Application Configuration -+SECRET_KEY=your_secret_key_for_sessions -+PORT=8080 -+ -+# Email Configuration (Optional) -+SMTP_SERVER=your_smtp_server -+SMTP_PORT=465 -+SMTP_USER=your_email@domain.com -+SMTP_PASSWORD=your_email_password - ``` --DATABASE= -+ -+### Running the Application -+ -+#### Development Mode -+ -+```bash -+go run main.go -+``` -+ -+The application will be available at http://localhost:8080. -+ -+#### Production Mode -+ -+```bash -+./rideaware-api - ``` --- Replace `` with the URI of your database (e.g., SQLite, PostgreSQL). -+ -+### API Endpoints -+ -+- `GET /health` - Health check endpoint -+- `POST /auth/signup` - User registration -+- `POST /auth/login` - User authentication -+- `POST /auth/logout` - User logout - - ### Running with Docker - --To run the application in a containerized environment, you can use the provided Dockerfile. -+To run the application in a containerized environment: - - 1. **Build the Docker Image**: - -@@ -72,14 +97,60 @@ docker build -t rideaware-api . - 2. **Run the Container** - - ```bash --docker run -d -p 5000:5000 --env-file .env rideaware-api -+docker run -d -p 8080:8080 --env-file .env rideaware-api -+``` -+ -+The application will be available at http://localhost:8080. -+ -+### Example Dockerfile -+ -+```dockerfile -+FROM golang:1.21-alpine AS builder -+ -+WORKDIR /app -+COPY go.mod go.sum ./ -+RUN go mod download -+ -+COPY . . -+RUN go build -o rideaware-api -+ -+FROM alpine:latest -+RUN apk --no-cache add ca-certificates -+WORKDIR /root/ -+ -+COPY --from=builder /app/rideaware-api . -+CMD ["./rideaware-api"] - ``` - --The application will be available at http://127.0.0.1:5000. -+### Database Migration -+ -+The application automatically runs database migrations on startup using GORM's AutoMigrate feature. This will create the necessary tables: -+ -+- `users` - User accounts -+- `user_profiles` - User profile information - - ### Running Tests - --To be added. -+To run tests: -+ -+```bash -+go test ./... -+``` -+ -+To run tests with coverage: -+ -+```bash -+go test -cover ./... -+``` -+ -+### Development -+ -+To add new features: -+ -+1. Create models in the `models/` directory -+2. Add business logic in the `services/` directory -+3. Define API routes in the `routes/` directory -+4. Register routes in `main.go` - - ## Contributing - -@@ -87,5 +158,4 @@ Contributions are welcome! Please create a pull request or open an issue for any - - ## License - --This project is licensed under the AGPL-3.0 License. -- -+This project is licensed under the AGPL-3.0 License. -\ No newline at end of file -diff --git a/migrations/README b/migrations/README -deleted file mode 100644 -index 0e04844..0000000 ---- a/migrations/README -+++ /dev/null -@@ -1 +0,0 @@ --Single-database configuration for Flask. -diff --git a/migrations/alembic.ini b/migrations/alembic.ini -deleted file mode 100644 -index ec9d45c..0000000 ---- a/migrations/alembic.ini -+++ /dev/null -@@ -1,50 +0,0 @@ --# A generic, single database configuration. -- --[alembic] --# template used to generate migration files --# file_template = %%(rev)s_%%(slug)s -- --# set to 'true' to run the environment during --# the 'revision' command, regardless of autogenerate --# revision_environment = false -- -- --# Logging configuration --[loggers] --keys = root,sqlalchemy,alembic,flask_migrate -- --[handlers] --keys = console -- --[formatters] --keys = generic -- --[logger_root] --level = WARN --handlers = console --qualname = -- --[logger_sqlalchemy] --level = WARN --handlers = --qualname = sqlalchemy.engine -- --[logger_alembic] --level = INFO --handlers = --qualname = alembic -- --[logger_flask_migrate] --level = INFO --handlers = --qualname = flask_migrate -- --[handler_console] --class = StreamHandler --args = (sys.stderr,) --level = NOTSET --formatter = generic -- --[formatter_generic] --format = %(levelname)-5.5s [%(name)s] %(message)s --datefmt = %H:%M:%S -diff --git a/migrations/env.py b/migrations/env.py -deleted file mode 100644 -index 4c97092..0000000 ---- a/migrations/env.py -+++ /dev/null -@@ -1,113 +0,0 @@ --import logging --from logging.config import fileConfig -- --from flask import current_app -- --from alembic import context -- --# this is the Alembic Config object, which provides --# access to the values within the .ini file in use. --config = context.config -- --# Interpret the config file for Python logging. --# This line sets up loggers basically. --fileConfig(config.config_file_name) --logger = logging.getLogger('alembic.env') -- -- --def get_engine(): -- try: -- # this works with Flask-SQLAlchemy<3 and Alchemical -- return current_app.extensions['migrate'].db.get_engine() -- except (TypeError, AttributeError): -- # this works with Flask-SQLAlchemy>=3 -- return current_app.extensions['migrate'].db.engine -- -- --def get_engine_url(): -- try: -- return get_engine().url.render_as_string(hide_password=False).replace( -- '%', '%%') -- except AttributeError: -- return str(get_engine().url).replace('%', '%%') -- -- --# add your model's MetaData object here --# for 'autogenerate' support --# from myapp import mymodel --# target_metadata = mymodel.Base.metadata --config.set_main_option('sqlalchemy.url', get_engine_url()) --target_db = current_app.extensions['migrate'].db -- --# other values from the config, defined by the needs of env.py, --# can be acquired: --# my_important_option = config.get_main_option("my_important_option") --# ... etc. -- -- --def get_metadata(): -- if hasattr(target_db, 'metadatas'): -- return target_db.metadatas[None] -- return target_db.metadata -- -- --def run_migrations_offline(): -- """Run migrations in 'offline' mode. -- -- This configures the context with just a URL -- and not an Engine, though an Engine is acceptable -- here as well. By skipping the Engine creation -- we don't even need a DBAPI to be available. -- -- Calls to context.execute() here emit the given string to the -- script output. -- -- """ -- url = config.get_main_option("sqlalchemy.url") -- context.configure( -- url=url, target_metadata=get_metadata(), literal_binds=True -- ) -- -- with context.begin_transaction(): -- context.run_migrations() -- -- --def run_migrations_online(): -- """Run migrations in 'online' mode. -- -- In this scenario we need to create an Engine -- and associate a connection with the context. -- -- """ -- -- # this callback is used to prevent an auto-migration from being generated -- # when there are no changes to the schema -- # reference: http://alembic.zzzcomputing.com/en/latest/cookbook.html -- def process_revision_directives(context, revision, directives): -- if getattr(config.cmd_opts, 'autogenerate', False): -- script = directives[0] -- if script.upgrade_ops.is_empty(): -- directives[:] = [] -- logger.info('No changes in schema detected.') -- -- conf_args = current_app.extensions['migrate'].configure_args -- if conf_args.get("process_revision_directives") is None: -- conf_args["process_revision_directives"] = process_revision_directives -- -- connectable = get_engine() -- -- with connectable.connect() as connection: -- context.configure( -- connection=connection, -- target_metadata=get_metadata(), -- **conf_args -- ) -- -- with context.begin_transaction(): -- context.run_migrations() -- -- --if context.is_offline_mode(): -- run_migrations_offline() --else: -- run_migrations_online() -diff --git a/migrations/script.py.mako b/migrations/script.py.mako -deleted file mode 100644 -index 2c01563..0000000 ---- a/migrations/script.py.mako -+++ /dev/null -@@ -1,24 +0,0 @@ --"""${message} -- --Revision ID: ${up_revision} --Revises: ${down_revision | comma,n} --Create Date: ${create_date} -- --""" --from alembic import op --import sqlalchemy as sa --${imports if imports else ""} -- --# revision identifiers, used by Alembic. --revision = ${repr(up_revision)} --down_revision = ${repr(down_revision)} --branch_labels = ${repr(branch_labels)} --depends_on = ${repr(depends_on)} -- -- --def upgrade(): -- ${upgrades if upgrades else "pass"} -- -- --def downgrade(): -- ${downgrades if downgrades else "pass"} -diff --git a/migrations/versions/0e07095d2961_initial_migration.py b/migrations/versions/0e07095d2961_initial_migration.py -deleted file mode 100644 -index 594c8d6..0000000 ---- a/migrations/versions/0e07095d2961_initial_migration.py -+++ /dev/null -@@ -1,99 +0,0 @@ --"""Initial migration -- --Revision ID: 0e07095d2961 --Revises: --Create Date: 2025-08-29 01:28:57.822103 -- --""" --from alembic import op --import sqlalchemy as sa --from sqlalchemy.dialects import postgresql -- --# revision identifiers, used by Alembic. --revision = '0e07095d2961' --down_revision = None --branch_labels = None --depends_on = None -- -- --def upgrade(): -- # ### commands auto generated by Alembic - please adjust! ### -- op.drop_table('admins') -- with op.batch_alter_table('subscribers', schema=None) as batch_op: -- batch_op.drop_index(batch_op.f('idx_subscribers_created_at')) -- batch_op.drop_index(batch_op.f('idx_subscribers_email')) -- batch_op.drop_index(batch_op.f('idx_subscribers_status')) -- -- op.drop_table('subscribers') -- op.drop_table('admin_users') -- op.drop_table('email_deliveries') -- with op.batch_alter_table('newsletters', schema=None) as batch_op: -- batch_op.drop_index(batch_op.f('idx_newsletters_sent_at')) -- -- op.drop_table('newsletters') -- # ### end Alembic commands ### -- -- --def downgrade(): -- # ### commands auto generated by Alembic - please adjust! ### -- op.create_table('newsletters', -- sa.Column('id', sa.INTEGER(), server_default=sa.text("nextval('newsletters_id_seq'::regclass)"), autoincrement=True, nullable=False), -- sa.Column('subject', sa.TEXT(), autoincrement=False, nullable=False), -- sa.Column('body', sa.TEXT(), autoincrement=False, nullable=False), -- sa.Column('sent_at', postgresql.TIMESTAMP(), server_default=sa.text('CURRENT_TIMESTAMP'), autoincrement=False, nullable=True), -- sa.Column('sent_by', sa.TEXT(), autoincrement=False, nullable=True), -- sa.Column('recipient_count', sa.INTEGER(), server_default=sa.text('0'), autoincrement=False, nullable=True), -- sa.Column('success_count', sa.INTEGER(), server_default=sa.text('0'), autoincrement=False, nullable=True), -- sa.Column('failure_count', sa.INTEGER(), server_default=sa.text('0'), autoincrement=False, nullable=True), -- sa.PrimaryKeyConstraint('id', name='newsletters_pkey'), -- postgresql_ignore_search_path=False -- ) -- with op.batch_alter_table('newsletters', schema=None) as batch_op: -- batch_op.create_index(batch_op.f('idx_newsletters_sent_at'), [sa.literal_column('sent_at DESC')], unique=False) -- -- op.create_table('email_deliveries', -- sa.Column('id', sa.INTEGER(), autoincrement=True, nullable=False), -- sa.Column('newsletter_id', sa.INTEGER(), autoincrement=False, nullable=True), -- sa.Column('email', sa.TEXT(), autoincrement=False, nullable=False), -- sa.Column('status', sa.TEXT(), autoincrement=False, nullable=True), -- sa.Column('sent_at', postgresql.TIMESTAMP(timezone=True), server_default=sa.text('CURRENT_TIMESTAMP'), autoincrement=False, nullable=True), -- sa.Column('error_message', sa.TEXT(), autoincrement=False, nullable=True), -- sa.CheckConstraint("status = ANY (ARRAY['sent'::text, 'failed'::text, 'bounced'::text])", name=op.f('email_deliveries_status_check')), -- sa.ForeignKeyConstraint(['newsletter_id'], ['newsletters.id'], name=op.f('email_deliveries_newsletter_id_fkey')), -- sa.PrimaryKeyConstraint('id', name=op.f('email_deliveries_pkey')) -- ) -- op.create_table('admin_users', -- sa.Column('id', sa.INTEGER(), autoincrement=True, nullable=False), -- sa.Column('username', sa.TEXT(), autoincrement=False, nullable=False), -- sa.Column('password', sa.TEXT(), autoincrement=False, nullable=False), -- sa.Column('created_at', postgresql.TIMESTAMP(timezone=True), server_default=sa.text('CURRENT_TIMESTAMP'), autoincrement=False, nullable=True), -- sa.Column('last_login', postgresql.TIMESTAMP(timezone=True), autoincrement=False, nullable=True), -- sa.Column('is_active', sa.BOOLEAN(), server_default=sa.text('true'), autoincrement=False, nullable=True), -- sa.PrimaryKeyConstraint('id', name=op.f('admin_users_pkey')), -- sa.UniqueConstraint('username', name=op.f('admin_users_username_key'), postgresql_include=[], postgresql_nulls_not_distinct=False) -- ) -- op.create_table('subscribers', -- sa.Column('id', sa.INTEGER(), autoincrement=True, nullable=False), -- sa.Column('email', sa.TEXT(), autoincrement=False, nullable=False), -- sa.Column('created_at', postgresql.TIMESTAMP(timezone=True), server_default=sa.text('CURRENT_TIMESTAMP'), autoincrement=False, nullable=True), -- sa.Column('subscribed_at', postgresql.TIMESTAMP(timezone=True), server_default=sa.text('CURRENT_TIMESTAMP'), autoincrement=False, nullable=True), -- sa.Column('status', sa.TEXT(), server_default=sa.text("'active'::text"), autoincrement=False, nullable=True), -- sa.Column('source', sa.TEXT(), server_default=sa.text("'manual'::text"), autoincrement=False, nullable=True), -- sa.CheckConstraint("status = ANY (ARRAY['active'::text, 'unsubscribed'::text])", name=op.f('subscribers_status_check')), -- sa.PrimaryKeyConstraint('id', name=op.f('subscribers_pkey')), -- sa.UniqueConstraint('email', name=op.f('subscribers_email_key'), postgresql_include=[], postgresql_nulls_not_distinct=False) -- ) -- with op.batch_alter_table('subscribers', schema=None) as batch_op: -- batch_op.create_index(batch_op.f('idx_subscribers_status'), ['status'], unique=False) -- batch_op.create_index(batch_op.f('idx_subscribers_email'), ['email'], unique=False) -- batch_op.create_index(batch_op.f('idx_subscribers_created_at'), [sa.literal_column('created_at DESC')], unique=False) -- -- op.create_table('admins', -- sa.Column('id', sa.INTEGER(), autoincrement=True, nullable=False), -- sa.Column('username', sa.VARCHAR(length=100), autoincrement=False, nullable=False), -- sa.Column('password_hash', sa.VARCHAR(length=255), autoincrement=False, nullable=False), -- sa.Column('created_at', postgresql.TIMESTAMP(), server_default=sa.text('CURRENT_TIMESTAMP'), autoincrement=False, nullable=True), -- sa.PrimaryKeyConstraint('id', name=op.f('admins_pkey')), -- sa.UniqueConstraint('username', name=op.f('admins_username_key'), postgresql_include=[], postgresql_nulls_not_distinct=False) -- ) -- # ### end Alembic commands ### -diff --git a/models/User/user.py b/models/User/user.py -deleted file mode 100644 -index 552796c..0000000 ---- a/models/User/user.py -+++ /dev/null -@@ -1,40 +0,0 @@ --from models.UserProfile.user_profile import UserProfile --from werkzeug.security import generate_password_hash, check_password_hash --from models import db --from sqlalchemy import event -- --class User(db.Model): -- __tablename__ = 'users' -- -- id = db.Column(db.Integer, primary_key=True) -- username = db.Column(db.String(80), unique=True, nullable=False) -- email = db.Column(db.String(120), unique=True, nullable=False) # Add email field -- _password = db.Column("password", db.String(255), nullable=False) -- -- profile = db.relationship('UserProfile', back_populates='user', uselist=False, cascade="all, delete-orphan") -- -- @property -- def password(self): -- return self._password -- -- @password.setter -- def password(self, raw_password): -- if not raw_password.startswith("pbkdf2:sha256:"): -- self._password = generate_password_hash(raw_password) -- else: -- self._password = raw_password -- -- def check_password(self, password): -- return check_password_hash(self._password, password) -- --@event.listens_for(User, 'after_insert') --def create_user_profile(mapper, connection, target): -- connection.execute( -- UserProfile.__table__.insert().values( -- user_id=target.id, -- first_name="", -- last_name="", -- bio="", -- profile_picture="" -- ) -- ) -\ No newline at end of file -diff --git a/models/UserProfile/user_profile.py b/models/UserProfile/user_profile.py -deleted file mode 100644 -index d3fa194..0000000 ---- a/models/UserProfile/user_profile.py -+++ /dev/null -@@ -1,13 +0,0 @@ --from models import db -- --class UserProfile(db.Model): -- __tablename__ = 'user_profiles' -- -- id = db.Column(db.Integer, primary_key=True) -- user_id = db.Column(db.Integer, db.ForeignKey('users.id'), nullable=False) -- first_name = db.Column(db.String(50), nullable=False, default="") -- last_name = db.Column(db.String(50), nullable=False, default="") -- bio = db.Column(db.Text, default="") -- profile_picture = db.Column(db.String(255), default="") -- -- user = db.relationship('User', back_populates='profile') -\ No newline at end of file -diff --git a/models/__init__.py b/models/__init__.py -deleted file mode 100644 -index 8dd3fe9..0000000 ---- a/models/__init__.py -+++ /dev/null -@@ -1,22 +0,0 @@ --import os --from flask_sqlalchemy import SQLAlchemy --from dotenv import load_dotenv --from urllib.parse import quote_plus -- --load_dotenv() -- --PG_USER = quote_plus(os.getenv("PG_USER", "postgres")) --PG_PASSWORD = quote_plus(os.getenv("PG_PASSWORD", "postgres")) --PG_HOST = os.getenv("PG_HOST", "localhost") --PG_PORT = os.getenv("PG_PORT", "5432") --PG_DATABASE = os.getenv("PG_DATABASE", "rideaware") -- --DATABASE_URI = f"postgresql+psycopg2://{PG_USER}:{PG_PASSWORD}@{PG_HOST}:{PG_PORT}/{PG_DATABASE}" -- --db = SQLAlchemy() -- --def init_db(app): -- """Initialize the SQLAlchemy app with the configuration.""" -- app.config['SQLALCHEMY_DATABASE_URI'] = DATABASE_URI -- app.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False -- db.init_app(app) -diff --git a/requirements.txt b/requirements.txt -deleted file mode 100644 -index 001e473..0000000 ---- a/requirements.txt -+++ /dev/null -@@ -1,8 +0,0 @@ --Flask --flask_bcrypt --flask_cors --flask_sqlalchemy --python-dotenv --werkzeug --psycopg2-binary --Flask-Migrate -\ No newline at end of file -diff --git a/routes/user_auth/auth.py b/routes/user_auth/auth.py -deleted file mode 100644 -index 899d7ba..0000000 ---- a/routes/user_auth/auth.py -+++ /dev/null -@@ -1,60 +0,0 @@ --from flask import Blueprint, request, jsonify, session --from services.UserService.user import UserService -- --auth_bp = Blueprint("auth", __name__, url_prefix="/api") --user_service = UserService() -- --@auth_bp.route("/signup", methods=["POST"]) --def signup(): -- data = request.get_json() -- if not data: -- return jsonify({"message": "No data provided"}), 400 -- -- required_fields = ['username', 'password'] -- for field in required_fields: -- if not data.get(field): -- return jsonify({"message": f"{field} is required"}), 400 -- -- try: -- new_user = user_service.create_user( -- username=data["username"], -- password=data["password"], -- email=data.get("email"), -- first_name=data.get("first_name"), -- last_name=data.get("last_name") -- ) -- -- return jsonify({ -- "message": "User created successfully", -- "username": new_user.username, -- "user_id": new_user.id -- }), 201 -- -- except ValueError as e: -- return jsonify({"message": str(e)}), 400 -- except Exception as e: -- # Log the error -- print(f"Signup error: {e}") -- return jsonify({"message": "Internal server error"}), 500 -- --@auth_bp.route("/login", methods=["POST"]) --def login(): -- data = request.get_json() -- username = data.get("username") -- password = data.get("password") -- print(f"Login attempt: username={username}, password={password}") -- try: -- user = user_service.verify_user(username, password) -- session["user_id"] = user.id -- return jsonify({"message": "Login successful", "user_id": user.id}), 200 -- except ValueError as e: -- print(f"Login failed: {str(e)}") -- return jsonify({"error": str(e)}), 401 -- except Exception as e: -- print(f"Login error: {e}") -- return jsonify({"error": "Internal server error"}), 500 -- --@auth_bp.route("/logout", methods=["POST"]) --def logout(): -- session.clear() -- return jsonify({"message": "Logout successful"}), 200 -\ No newline at end of file -diff --git a/scripts/migrate.sh b/scripts/migrate.sh -deleted file mode 100644 -index 405f399..0000000 ---- a/scripts/migrate.sh -+++ /dev/null -@@ -1,8 +0,0 @@ --#!/bin/bash --set -e -- --echo "Running database migrations..." --flask db upgrade -- --echo "Starting application..." --exec "$@" -\ No newline at end of file -diff --git a/server.py b/server.py -deleted file mode 100644 -index 5800353..0000000 ---- a/server.py -+++ /dev/null -@@ -1,33 +0,0 @@ --import os --from flask import Flask --from flask_cors import CORS --from dotenv import load_dotenv --from flask_migrate import Migrate --from flask.cli import FlaskGroup -- --from models import db, init_db --from routes.user_auth import auth -- --load_dotenv() -- --app = Flask(__name__) --app.config["SECRET_KEY"] = os.getenv("SECRET_KEY") --app.config["SQLALCHEMY_DATABASE_URI"] = os.getenv("DATABASE") --app.config["SQLALCHEMY_TRACK_MODIFICATIONS"] = False -- --CORS(app) -- --init_db(app) --migrate = Migrate(app, db) --app.register_blueprint(auth.auth_bp) -- -- --@app.route("/health") --def health_check(): -- """Health check endpoint.""" -- return "OK", 200 -- --cli = FlaskGroup(app) -- --if __name__ == "__main__": -- cli() -\ No newline at end of file -diff --git a/services/UserService/user.py b/services/UserService/user.py -deleted file mode 100644 -index 6f1c030..0000000 ---- a/services/UserService/user.py -+++ /dev/null -@@ -1,60 +0,0 @@ --from models.User.user import User --from models.UserProfile.user_profile import UserProfile --from models import db --import re -- --class UserService: -- def create_user(self, username, password, email=None, first_name=None, last_name=None): -- if not username or not password: -- raise ValueError("Username and password are required") -- -- if email: -- email_regex = r'^[a-zA-Z0-9._%+-]+@[a-zA-Z0-9.-]+\.[a-zA-Z]{2,}$' -- if not re.match(email_regex, email): -- raise ValueError("Invalid email format") -- -- existing_user = User.query.filter( -- (User.username == username) | (User.email == email) -- ).first() -- -- if existing_user: -- if existing_user.username == username: -- raise ValueError("Username already exists") -- else: -- raise ValueError("Email already exists") -- -- if len(password) < 8: -- raise ValueError("Password must be at least 8 characters long") -- -- try: -- new_user = User( -- username=username, -- email=email or "", -- password=password -- ) -- -- db.session.add(new_user) -- db.session.flush() -- -- user_profile = UserProfile( -- user_id=new_user.id, -- first_name=first_name or "", -- last_name=last_name or "", -- bio="", -- profile_picture="" -- ) -- -- db.session.add(user_profile) -- db.session.commit() -- -- return new_user -- -- except Exception as e: -- db.session.rollback() -- raise Exception(f"Error creating user: {str(e)}") -- -- def verify_user(self, username, password): -- user = User.query.filter_by(username=username).first() -- if not user or not user.check_password(password): -- raise ValueError("Invalid username or password") -- return user -\ No newline at end of file