diff --git a/diff.txt b/diff.txt new file mode 100644 index 0000000..7394524 --- /dev/null +++ b/diff.txt @@ -0,0 +1,915 @@ +diff --git a/.dockerignore b/.dockerignore +index d9b625e..3f2b844 100644 +--- a/.dockerignore ++++ b/.dockerignore +@@ -1,9 +1,13 @@ ++.env + .git +-__pycache__/ +-*.py[cod] ++.gitignore ++README.md ++Dockerfile ++.dockerignore ++.air.toml ++tmp/ ++.vscode/ ++.idea/ + *.log +-!.env +-venv/ +-.venv/ +-dist/ +-build/ +\ No newline at end of file ++coverage.out ++rideaware-api +\ No newline at end of file +diff --git a/Dockerfile b/Dockerfile +index c72dff0..7aa361b 100644 +--- a/Dockerfile ++++ b/Dockerfile +@@ -1,53 +1,52 @@ +-FROM python:3.10-slim AS builder +- +-ENV PYTHONDONTWRITEBYTECODE=1 \ +- PYTHONUNBUFFERED=1 \ +- PIP_NO_CACHE_DIR=1 ++# Build stage ++FROM golang:1.21-alpine AS builder + ++# Set working directory + WORKDIR /app + +-RUN apt-get update && apt-get install -y --no-install-recommends \ +- build-essential gcc \ +- && rm -rf /var/lib/apt/lists/* ++# Install git (needed for some Go modules) ++RUN apk add --no-cache git + +-COPY requirements.txt . ++# Copy go mod files ++COPY go.mod go.sum ./ + +-RUN python -m pip install --upgrade pip && \ +- pip wheel --no-deps -r requirements.txt -w /wheels && \ +- pip wheel --no-deps gunicorn -w /wheels ++# Download dependencies ++RUN go mod download + +-FROM python:3.10-slim AS runtime ++# Copy source code ++COPY . . + +-ENV PYTHONDONTWRITEBYTECODE=1 \ +- PYTHONUNBUFFERED=1 \ +- PIP_NO_CACHE_DIR=1 \ +- PORT=5000 \ +- WSGI_MODULE=server:app \ +- GUNICORN_WORKERS=2 \ +- GUNICORN_THREADS=4 \ +- GUNICORN_TIMEOUT=60 \ +- FLASK_APP=server.py ++# Build the application ++RUN CGO_ENABLED=0 GOOS=linux go build -a -installsuffix cgo -o rideaware-api . + +-WORKDIR /app ++# Production stage ++FROM alpine:latest + +-RUN groupadd -g 10001 app && useradd -m -u 10001 -g app app ++# Install ca-certificates for HTTPS requests and timezone data ++RUN apk --no-cache add ca-certificates tzdata + +-COPY --from=builder /wheels /wheels +-RUN pip install --no-cache-dir /wheels/* && rm -rf /wheels ++# Create non-root user ++RUN addgroup -g 1001 -S appgroup && \ ++ adduser -u 1001 -S appuser -G appgroup + +-# Install python-dotenv if not already in requirements.txt +-RUN pip install python-dotenv ++# Set working directory ++WORKDIR /home/appuser + +-USER app ++# Copy binary from builder stage ++COPY --from=builder /app/rideaware-api . + +-COPY --chown=app:app . . ++# Change ownership to non-root user ++RUN chown -R appuser:appgroup /home/appuser + +-# Copy .env file specifically +-COPY --chown=app:app .env .env ++# Switch to non-root user ++USER appuser + +-EXPOSE 5000 ++# Expose port ++EXPOSE 8080 + +-HEALTHCHECK --interval=30s --timeout=3s --start-period=10s --retries=3 \ +- CMD python -c "import os,socket; s=socket.socket(); s.settimeout(2); s.connect(('127.0.0.1', int(os.getenv('PORT', '5000')))); s.close()" ++# Health check ++HEALTHCHECK --interval=30s --timeout=3s --start-period=5s --retries=3 \ ++ CMD wget --no-verbose --tries=1 --spider http://localhost:8080/health || exit 1 + +-CMD ["sh", "-c", "exec gunicorn $WSGI_MODULE --bind=0.0.0.0:$PORT --workers=$GUNICORN_WORKERS --threads=$GUNICORN_THREADS --timeout=$GUNICORN_TIMEOUT --access-logfile=- --error-logfile=- --keep-alive=5"] +\ No newline at end of file ++# Run the application ++CMD ["./rideaware-api"] +\ No newline at end of file +diff --git a/README.md b/README.md +index cda9868..65ce1b9 100644 +--- a/README.md ++++ b/README.md +@@ -15,53 +15,78 @@ Whether you're building a structured training plan, analyzing ride data, or comp + Ensure you have the following installed on your system: + + - Docker +-- Python 3.10 or later +-- pip ++- Go 1.21 or later ++- PostgreSQL (for local development, optional) + + ### Setting Up the Project + + 1. **Clone the Repository** + + ```bash +- git clone https://github.com/VeloInnovate/rideaware-api.git ++ git clone https://github.com/rideaware/rideaware-api.git + cd rideaware-api + ``` + +-2. **Create a Virtual Environment** +- It is recommended to use a Python virtual environment to isolate dependencies. ++2. **Install Go Dependencies** + + ```bash +- python3 -m venv .venv ++ go mod tidy + ``` + +-3. **Activate the Virtual Environment** +- - On Linux/Mac: +- ```bash +- source .venv/bin/activate +- ``` +- - On Windows: +- ```cmd +- .venv\Scripts\activate +- ``` +- +-4. **Install Requirements** +- Install the required Python packages using pip: ++3. **Build the Application** + ```bash +- pip install -r requirements.txt ++ go build -o rideaware-api + ``` + + ### Configuration + + The application uses environment variables for configuration. Create a `.env` file in the root directory and define the following variables: + ++```env ++# Database Configuration ++PG_HOST=your_postgres_host ++PG_PORT=5432 ++PG_DATABASE=rideaware ++PG_USER=your_postgres_user ++PG_PASSWORD=your_postgres_password ++ ++# Application Configuration ++SECRET_KEY=your_secret_key_for_sessions ++PORT=8080 ++ ++# Email Configuration (Optional) ++SMTP_SERVER=your_smtp_server ++SMTP_PORT=465 ++SMTP_USER=your_email@domain.com ++SMTP_PASSWORD=your_email_password + ``` +-DATABASE= ++ ++### Running the Application ++ ++#### Development Mode ++ ++```bash ++go run main.go ++``` ++ ++The application will be available at http://localhost:8080. ++ ++#### Production Mode ++ ++```bash ++./rideaware-api + ``` +-- Replace `` with the URI of your database (e.g., SQLite, PostgreSQL). ++ ++### API Endpoints ++ ++- `GET /health` - Health check endpoint ++- `POST /auth/signup` - User registration ++- `POST /auth/login` - User authentication ++- `POST /auth/logout` - User logout + + ### Running with Docker + +-To run the application in a containerized environment, you can use the provided Dockerfile. ++To run the application in a containerized environment: + + 1. **Build the Docker Image**: + +@@ -72,14 +97,60 @@ docker build -t rideaware-api . + 2. **Run the Container** + + ```bash +-docker run -d -p 5000:5000 --env-file .env rideaware-api ++docker run -d -p 8080:8080 --env-file .env rideaware-api ++``` ++ ++The application will be available at http://localhost:8080. ++ ++### Example Dockerfile ++ ++```dockerfile ++FROM golang:1.21-alpine AS builder ++ ++WORKDIR /app ++COPY go.mod go.sum ./ ++RUN go mod download ++ ++COPY . . ++RUN go build -o rideaware-api ++ ++FROM alpine:latest ++RUN apk --no-cache add ca-certificates ++WORKDIR /root/ ++ ++COPY --from=builder /app/rideaware-api . ++CMD ["./rideaware-api"] + ``` + +-The application will be available at http://127.0.0.1:5000. ++### Database Migration ++ ++The application automatically runs database migrations on startup using GORM's AutoMigrate feature. This will create the necessary tables: ++ ++- `users` - User accounts ++- `user_profiles` - User profile information + + ### Running Tests + +-To be added. ++To run tests: ++ ++```bash ++go test ./... ++``` ++ ++To run tests with coverage: ++ ++```bash ++go test -cover ./... ++``` ++ ++### Development ++ ++To add new features: ++ ++1. Create models in the `models/` directory ++2. Add business logic in the `services/` directory ++3. Define API routes in the `routes/` directory ++4. Register routes in `main.go` + + ## Contributing + +@@ -87,5 +158,4 @@ Contributions are welcome! Please create a pull request or open an issue for any + + ## License + +-This project is licensed under the AGPL-3.0 License. +- ++This project is licensed under the AGPL-3.0 License. +\ No newline at end of file +diff --git a/migrations/README b/migrations/README +deleted file mode 100644 +index 0e04844..0000000 +--- a/migrations/README ++++ /dev/null +@@ -1 +0,0 @@ +-Single-database configuration for Flask. +diff --git a/migrations/alembic.ini b/migrations/alembic.ini +deleted file mode 100644 +index ec9d45c..0000000 +--- a/migrations/alembic.ini ++++ /dev/null +@@ -1,50 +0,0 @@ +-# A generic, single database configuration. +- +-[alembic] +-# template used to generate migration files +-# file_template = %%(rev)s_%%(slug)s +- +-# set to 'true' to run the environment during +-# the 'revision' command, regardless of autogenerate +-# revision_environment = false +- +- +-# Logging configuration +-[loggers] +-keys = root,sqlalchemy,alembic,flask_migrate +- +-[handlers] +-keys = console +- +-[formatters] +-keys = generic +- +-[logger_root] +-level = WARN +-handlers = console +-qualname = +- +-[logger_sqlalchemy] +-level = WARN +-handlers = +-qualname = sqlalchemy.engine +- +-[logger_alembic] +-level = INFO +-handlers = +-qualname = alembic +- +-[logger_flask_migrate] +-level = INFO +-handlers = +-qualname = flask_migrate +- +-[handler_console] +-class = StreamHandler +-args = (sys.stderr,) +-level = NOTSET +-formatter = generic +- +-[formatter_generic] +-format = %(levelname)-5.5s [%(name)s] %(message)s +-datefmt = %H:%M:%S +diff --git a/migrations/env.py b/migrations/env.py +deleted file mode 100644 +index 4c97092..0000000 +--- a/migrations/env.py ++++ /dev/null +@@ -1,113 +0,0 @@ +-import logging +-from logging.config import fileConfig +- +-from flask import current_app +- +-from alembic import context +- +-# this is the Alembic Config object, which provides +-# access to the values within the .ini file in use. +-config = context.config +- +-# Interpret the config file for Python logging. +-# This line sets up loggers basically. +-fileConfig(config.config_file_name) +-logger = logging.getLogger('alembic.env') +- +- +-def get_engine(): +- try: +- # this works with Flask-SQLAlchemy<3 and Alchemical +- return current_app.extensions['migrate'].db.get_engine() +- except (TypeError, AttributeError): +- # this works with Flask-SQLAlchemy>=3 +- return current_app.extensions['migrate'].db.engine +- +- +-def get_engine_url(): +- try: +- return get_engine().url.render_as_string(hide_password=False).replace( +- '%', '%%') +- except AttributeError: +- return str(get_engine().url).replace('%', '%%') +- +- +-# add your model's MetaData object here +-# for 'autogenerate' support +-# from myapp import mymodel +-# target_metadata = mymodel.Base.metadata +-config.set_main_option('sqlalchemy.url', get_engine_url()) +-target_db = current_app.extensions['migrate'].db +- +-# other values from the config, defined by the needs of env.py, +-# can be acquired: +-# my_important_option = config.get_main_option("my_important_option") +-# ... etc. +- +- +-def get_metadata(): +- if hasattr(target_db, 'metadatas'): +- return target_db.metadatas[None] +- return target_db.metadata +- +- +-def run_migrations_offline(): +- """Run migrations in 'offline' mode. +- +- This configures the context with just a URL +- and not an Engine, though an Engine is acceptable +- here as well. By skipping the Engine creation +- we don't even need a DBAPI to be available. +- +- Calls to context.execute() here emit the given string to the +- script output. +- +- """ +- url = config.get_main_option("sqlalchemy.url") +- context.configure( +- url=url, target_metadata=get_metadata(), literal_binds=True +- ) +- +- with context.begin_transaction(): +- context.run_migrations() +- +- +-def run_migrations_online(): +- """Run migrations in 'online' mode. +- +- In this scenario we need to create an Engine +- and associate a connection with the context. +- +- """ +- +- # this callback is used to prevent an auto-migration from being generated +- # when there are no changes to the schema +- # reference: http://alembic.zzzcomputing.com/en/latest/cookbook.html +- def process_revision_directives(context, revision, directives): +- if getattr(config.cmd_opts, 'autogenerate', False): +- script = directives[0] +- if script.upgrade_ops.is_empty(): +- directives[:] = [] +- logger.info('No changes in schema detected.') +- +- conf_args = current_app.extensions['migrate'].configure_args +- if conf_args.get("process_revision_directives") is None: +- conf_args["process_revision_directives"] = process_revision_directives +- +- connectable = get_engine() +- +- with connectable.connect() as connection: +- context.configure( +- connection=connection, +- target_metadata=get_metadata(), +- **conf_args +- ) +- +- with context.begin_transaction(): +- context.run_migrations() +- +- +-if context.is_offline_mode(): +- run_migrations_offline() +-else: +- run_migrations_online() +diff --git a/migrations/script.py.mako b/migrations/script.py.mako +deleted file mode 100644 +index 2c01563..0000000 +--- a/migrations/script.py.mako ++++ /dev/null +@@ -1,24 +0,0 @@ +-"""${message} +- +-Revision ID: ${up_revision} +-Revises: ${down_revision | comma,n} +-Create Date: ${create_date} +- +-""" +-from alembic import op +-import sqlalchemy as sa +-${imports if imports else ""} +- +-# revision identifiers, used by Alembic. +-revision = ${repr(up_revision)} +-down_revision = ${repr(down_revision)} +-branch_labels = ${repr(branch_labels)} +-depends_on = ${repr(depends_on)} +- +- +-def upgrade(): +- ${upgrades if upgrades else "pass"} +- +- +-def downgrade(): +- ${downgrades if downgrades else "pass"} +diff --git a/migrations/versions/0e07095d2961_initial_migration.py b/migrations/versions/0e07095d2961_initial_migration.py +deleted file mode 100644 +index 594c8d6..0000000 +--- a/migrations/versions/0e07095d2961_initial_migration.py ++++ /dev/null +@@ -1,99 +0,0 @@ +-"""Initial migration +- +-Revision ID: 0e07095d2961 +-Revises: +-Create Date: 2025-08-29 01:28:57.822103 +- +-""" +-from alembic import op +-import sqlalchemy as sa +-from sqlalchemy.dialects import postgresql +- +-# revision identifiers, used by Alembic. +-revision = '0e07095d2961' +-down_revision = None +-branch_labels = None +-depends_on = None +- +- +-def upgrade(): +- # ### commands auto generated by Alembic - please adjust! ### +- op.drop_table('admins') +- with op.batch_alter_table('subscribers', schema=None) as batch_op: +- batch_op.drop_index(batch_op.f('idx_subscribers_created_at')) +- batch_op.drop_index(batch_op.f('idx_subscribers_email')) +- batch_op.drop_index(batch_op.f('idx_subscribers_status')) +- +- op.drop_table('subscribers') +- op.drop_table('admin_users') +- op.drop_table('email_deliveries') +- with op.batch_alter_table('newsletters', schema=None) as batch_op: +- batch_op.drop_index(batch_op.f('idx_newsletters_sent_at')) +- +- op.drop_table('newsletters') +- # ### end Alembic commands ### +- +- +-def downgrade(): +- # ### commands auto generated by Alembic - please adjust! ### +- op.create_table('newsletters', +- sa.Column('id', sa.INTEGER(), server_default=sa.text("nextval('newsletters_id_seq'::regclass)"), autoincrement=True, nullable=False), +- sa.Column('subject', sa.TEXT(), autoincrement=False, nullable=False), +- sa.Column('body', sa.TEXT(), autoincrement=False, nullable=False), +- sa.Column('sent_at', postgresql.TIMESTAMP(), server_default=sa.text('CURRENT_TIMESTAMP'), autoincrement=False, nullable=True), +- sa.Column('sent_by', sa.TEXT(), autoincrement=False, nullable=True), +- sa.Column('recipient_count', sa.INTEGER(), server_default=sa.text('0'), autoincrement=False, nullable=True), +- sa.Column('success_count', sa.INTEGER(), server_default=sa.text('0'), autoincrement=False, nullable=True), +- sa.Column('failure_count', sa.INTEGER(), server_default=sa.text('0'), autoincrement=False, nullable=True), +- sa.PrimaryKeyConstraint('id', name='newsletters_pkey'), +- postgresql_ignore_search_path=False +- ) +- with op.batch_alter_table('newsletters', schema=None) as batch_op: +- batch_op.create_index(batch_op.f('idx_newsletters_sent_at'), [sa.literal_column('sent_at DESC')], unique=False) +- +- op.create_table('email_deliveries', +- sa.Column('id', sa.INTEGER(), autoincrement=True, nullable=False), +- sa.Column('newsletter_id', sa.INTEGER(), autoincrement=False, nullable=True), +- sa.Column('email', sa.TEXT(), autoincrement=False, nullable=False), +- sa.Column('status', sa.TEXT(), autoincrement=False, nullable=True), +- sa.Column('sent_at', postgresql.TIMESTAMP(timezone=True), server_default=sa.text('CURRENT_TIMESTAMP'), autoincrement=False, nullable=True), +- sa.Column('error_message', sa.TEXT(), autoincrement=False, nullable=True), +- sa.CheckConstraint("status = ANY (ARRAY['sent'::text, 'failed'::text, 'bounced'::text])", name=op.f('email_deliveries_status_check')), +- sa.ForeignKeyConstraint(['newsletter_id'], ['newsletters.id'], name=op.f('email_deliveries_newsletter_id_fkey')), +- sa.PrimaryKeyConstraint('id', name=op.f('email_deliveries_pkey')) +- ) +- op.create_table('admin_users', +- sa.Column('id', sa.INTEGER(), autoincrement=True, nullable=False), +- sa.Column('username', sa.TEXT(), autoincrement=False, nullable=False), +- sa.Column('password', sa.TEXT(), autoincrement=False, nullable=False), +- sa.Column('created_at', postgresql.TIMESTAMP(timezone=True), server_default=sa.text('CURRENT_TIMESTAMP'), autoincrement=False, nullable=True), +- sa.Column('last_login', postgresql.TIMESTAMP(timezone=True), autoincrement=False, nullable=True), +- sa.Column('is_active', sa.BOOLEAN(), server_default=sa.text('true'), autoincrement=False, nullable=True), +- sa.PrimaryKeyConstraint('id', name=op.f('admin_users_pkey')), +- sa.UniqueConstraint('username', name=op.f('admin_users_username_key'), postgresql_include=[], postgresql_nulls_not_distinct=False) +- ) +- op.create_table('subscribers', +- sa.Column('id', sa.INTEGER(), autoincrement=True, nullable=False), +- sa.Column('email', sa.TEXT(), autoincrement=False, nullable=False), +- sa.Column('created_at', postgresql.TIMESTAMP(timezone=True), server_default=sa.text('CURRENT_TIMESTAMP'), autoincrement=False, nullable=True), +- sa.Column('subscribed_at', postgresql.TIMESTAMP(timezone=True), server_default=sa.text('CURRENT_TIMESTAMP'), autoincrement=False, nullable=True), +- sa.Column('status', sa.TEXT(), server_default=sa.text("'active'::text"), autoincrement=False, nullable=True), +- sa.Column('source', sa.TEXT(), server_default=sa.text("'manual'::text"), autoincrement=False, nullable=True), +- sa.CheckConstraint("status = ANY (ARRAY['active'::text, 'unsubscribed'::text])", name=op.f('subscribers_status_check')), +- sa.PrimaryKeyConstraint('id', name=op.f('subscribers_pkey')), +- sa.UniqueConstraint('email', name=op.f('subscribers_email_key'), postgresql_include=[], postgresql_nulls_not_distinct=False) +- ) +- with op.batch_alter_table('subscribers', schema=None) as batch_op: +- batch_op.create_index(batch_op.f('idx_subscribers_status'), ['status'], unique=False) +- batch_op.create_index(batch_op.f('idx_subscribers_email'), ['email'], unique=False) +- batch_op.create_index(batch_op.f('idx_subscribers_created_at'), [sa.literal_column('created_at DESC')], unique=False) +- +- op.create_table('admins', +- sa.Column('id', sa.INTEGER(), autoincrement=True, nullable=False), +- sa.Column('username', sa.VARCHAR(length=100), autoincrement=False, nullable=False), +- sa.Column('password_hash', sa.VARCHAR(length=255), autoincrement=False, nullable=False), +- sa.Column('created_at', postgresql.TIMESTAMP(), server_default=sa.text('CURRENT_TIMESTAMP'), autoincrement=False, nullable=True), +- sa.PrimaryKeyConstraint('id', name=op.f('admins_pkey')), +- sa.UniqueConstraint('username', name=op.f('admins_username_key'), postgresql_include=[], postgresql_nulls_not_distinct=False) +- ) +- # ### end Alembic commands ### +diff --git a/models/User/user.py b/models/User/user.py +deleted file mode 100644 +index 552796c..0000000 +--- a/models/User/user.py ++++ /dev/null +@@ -1,40 +0,0 @@ +-from models.UserProfile.user_profile import UserProfile +-from werkzeug.security import generate_password_hash, check_password_hash +-from models import db +-from sqlalchemy import event +- +-class User(db.Model): +- __tablename__ = 'users' +- +- id = db.Column(db.Integer, primary_key=True) +- username = db.Column(db.String(80), unique=True, nullable=False) +- email = db.Column(db.String(120), unique=True, nullable=False) # Add email field +- _password = db.Column("password", db.String(255), nullable=False) +- +- profile = db.relationship('UserProfile', back_populates='user', uselist=False, cascade="all, delete-orphan") +- +- @property +- def password(self): +- return self._password +- +- @password.setter +- def password(self, raw_password): +- if not raw_password.startswith("pbkdf2:sha256:"): +- self._password = generate_password_hash(raw_password) +- else: +- self._password = raw_password +- +- def check_password(self, password): +- return check_password_hash(self._password, password) +- +-@event.listens_for(User, 'after_insert') +-def create_user_profile(mapper, connection, target): +- connection.execute( +- UserProfile.__table__.insert().values( +- user_id=target.id, +- first_name="", +- last_name="", +- bio="", +- profile_picture="" +- ) +- ) +\ No newline at end of file +diff --git a/models/UserProfile/user_profile.py b/models/UserProfile/user_profile.py +deleted file mode 100644 +index d3fa194..0000000 +--- a/models/UserProfile/user_profile.py ++++ /dev/null +@@ -1,13 +0,0 @@ +-from models import db +- +-class UserProfile(db.Model): +- __tablename__ = 'user_profiles' +- +- id = db.Column(db.Integer, primary_key=True) +- user_id = db.Column(db.Integer, db.ForeignKey('users.id'), nullable=False) +- first_name = db.Column(db.String(50), nullable=False, default="") +- last_name = db.Column(db.String(50), nullable=False, default="") +- bio = db.Column(db.Text, default="") +- profile_picture = db.Column(db.String(255), default="") +- +- user = db.relationship('User', back_populates='profile') +\ No newline at end of file +diff --git a/models/__init__.py b/models/__init__.py +deleted file mode 100644 +index 8dd3fe9..0000000 +--- a/models/__init__.py ++++ /dev/null +@@ -1,22 +0,0 @@ +-import os +-from flask_sqlalchemy import SQLAlchemy +-from dotenv import load_dotenv +-from urllib.parse import quote_plus +- +-load_dotenv() +- +-PG_USER = quote_plus(os.getenv("PG_USER", "postgres")) +-PG_PASSWORD = quote_plus(os.getenv("PG_PASSWORD", "postgres")) +-PG_HOST = os.getenv("PG_HOST", "localhost") +-PG_PORT = os.getenv("PG_PORT", "5432") +-PG_DATABASE = os.getenv("PG_DATABASE", "rideaware") +- +-DATABASE_URI = f"postgresql+psycopg2://{PG_USER}:{PG_PASSWORD}@{PG_HOST}:{PG_PORT}/{PG_DATABASE}" +- +-db = SQLAlchemy() +- +-def init_db(app): +- """Initialize the SQLAlchemy app with the configuration.""" +- app.config['SQLALCHEMY_DATABASE_URI'] = DATABASE_URI +- app.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False +- db.init_app(app) +diff --git a/requirements.txt b/requirements.txt +deleted file mode 100644 +index 001e473..0000000 +--- a/requirements.txt ++++ /dev/null +@@ -1,8 +0,0 @@ +-Flask +-flask_bcrypt +-flask_cors +-flask_sqlalchemy +-python-dotenv +-werkzeug +-psycopg2-binary +-Flask-Migrate +\ No newline at end of file +diff --git a/routes/user_auth/auth.py b/routes/user_auth/auth.py +deleted file mode 100644 +index 899d7ba..0000000 +--- a/routes/user_auth/auth.py ++++ /dev/null +@@ -1,60 +0,0 @@ +-from flask import Blueprint, request, jsonify, session +-from services.UserService.user import UserService +- +-auth_bp = Blueprint("auth", __name__, url_prefix="/api") +-user_service = UserService() +- +-@auth_bp.route("/signup", methods=["POST"]) +-def signup(): +- data = request.get_json() +- if not data: +- return jsonify({"message": "No data provided"}), 400 +- +- required_fields = ['username', 'password'] +- for field in required_fields: +- if not data.get(field): +- return jsonify({"message": f"{field} is required"}), 400 +- +- try: +- new_user = user_service.create_user( +- username=data["username"], +- password=data["password"], +- email=data.get("email"), +- first_name=data.get("first_name"), +- last_name=data.get("last_name") +- ) +- +- return jsonify({ +- "message": "User created successfully", +- "username": new_user.username, +- "user_id": new_user.id +- }), 201 +- +- except ValueError as e: +- return jsonify({"message": str(e)}), 400 +- except Exception as e: +- # Log the error +- print(f"Signup error: {e}") +- return jsonify({"message": "Internal server error"}), 500 +- +-@auth_bp.route("/login", methods=["POST"]) +-def login(): +- data = request.get_json() +- username = data.get("username") +- password = data.get("password") +- print(f"Login attempt: username={username}, password={password}") +- try: +- user = user_service.verify_user(username, password) +- session["user_id"] = user.id +- return jsonify({"message": "Login successful", "user_id": user.id}), 200 +- except ValueError as e: +- print(f"Login failed: {str(e)}") +- return jsonify({"error": str(e)}), 401 +- except Exception as e: +- print(f"Login error: {e}") +- return jsonify({"error": "Internal server error"}), 500 +- +-@auth_bp.route("/logout", methods=["POST"]) +-def logout(): +- session.clear() +- return jsonify({"message": "Logout successful"}), 200 +\ No newline at end of file +diff --git a/scripts/migrate.sh b/scripts/migrate.sh +deleted file mode 100644 +index 405f399..0000000 +--- a/scripts/migrate.sh ++++ /dev/null +@@ -1,8 +0,0 @@ +-#!/bin/bash +-set -e +- +-echo "Running database migrations..." +-flask db upgrade +- +-echo "Starting application..." +-exec "$@" +\ No newline at end of file +diff --git a/server.py b/server.py +deleted file mode 100644 +index 5800353..0000000 +--- a/server.py ++++ /dev/null +@@ -1,33 +0,0 @@ +-import os +-from flask import Flask +-from flask_cors import CORS +-from dotenv import load_dotenv +-from flask_migrate import Migrate +-from flask.cli import FlaskGroup +- +-from models import db, init_db +-from routes.user_auth import auth +- +-load_dotenv() +- +-app = Flask(__name__) +-app.config["SECRET_KEY"] = os.getenv("SECRET_KEY") +-app.config["SQLALCHEMY_DATABASE_URI"] = os.getenv("DATABASE") +-app.config["SQLALCHEMY_TRACK_MODIFICATIONS"] = False +- +-CORS(app) +- +-init_db(app) +-migrate = Migrate(app, db) +-app.register_blueprint(auth.auth_bp) +- +- +-@app.route("/health") +-def health_check(): +- """Health check endpoint.""" +- return "OK", 200 +- +-cli = FlaskGroup(app) +- +-if __name__ == "__main__": +- cli() +\ No newline at end of file +diff --git a/services/UserService/user.py b/services/UserService/user.py +deleted file mode 100644 +index 6f1c030..0000000 +--- a/services/UserService/user.py ++++ /dev/null +@@ -1,60 +0,0 @@ +-from models.User.user import User +-from models.UserProfile.user_profile import UserProfile +-from models import db +-import re +- +-class UserService: +- def create_user(self, username, password, email=None, first_name=None, last_name=None): +- if not username or not password: +- raise ValueError("Username and password are required") +- +- if email: +- email_regex = r'^[a-zA-Z0-9._%+-]+@[a-zA-Z0-9.-]+\.[a-zA-Z]{2,}$' +- if not re.match(email_regex, email): +- raise ValueError("Invalid email format") +- +- existing_user = User.query.filter( +- (User.username == username) | (User.email == email) +- ).first() +- +- if existing_user: +- if existing_user.username == username: +- raise ValueError("Username already exists") +- else: +- raise ValueError("Email already exists") +- +- if len(password) < 8: +- raise ValueError("Password must be at least 8 characters long") +- +- try: +- new_user = User( +- username=username, +- email=email or "", +- password=password +- ) +- +- db.session.add(new_user) +- db.session.flush() +- +- user_profile = UserProfile( +- user_id=new_user.id, +- first_name=first_name or "", +- last_name=last_name or "", +- bio="", +- profile_picture="" +- ) +- +- db.session.add(user_profile) +- db.session.commit() +- +- return new_user +- +- except Exception as e: +- db.session.rollback() +- raise Exception(f"Error creating user: {str(e)}") +- +- def verify_user(self, username, password): +- user = User.query.filter_by(username=username).first() +- if not user or not user.check_password(password): +- raise ValueError("Invalid username or password") +- return user +\ No newline at end of file