top of page

Things I have been working on

Projects

Data Specialist for Production Plant

  • Centralized data from multiple sources into various databases using ETL (Alteryx) and SQL.

  • Designed and implemented data pipelines to automate the extraction, transformation, and loading (ETL) processes.

  • Utilized Alteryx for data preparation, blending, and analytics, ensuring data integrity and consistency.

  • Employed SQL queries for efficient data extraction, joining, and aggregation across different databases.

  • Developed dashboards and reports using Power BI to visualize production metrics and key performance indicators (KPIs).

  • Integrated data from SAP, MES (Manufacturing Execution Systems), and other operational databases into a centralized data warehouse.

  • Conducted data quality checks and validation to ensure accurate and reliable data for analysis.

  • Streamlined data workflows, reducing manual data handling, which resulted in saving up to 1000 human hours.

  • Provided data-driven insights to optimize production processes, enhance efficiency, and support decision-making.

  • Collaborated with cross-functional teams, including IT, operations, and finance, to align data strategies with business goals.

  • Created detailed Power BI dashboards that included interactive visualizations, allowing stakeholders to explore data and derive insights dynamically.

  • Utilized advanced Power BI features such as DAX (Data Analysis Expressions) and Power Query for complex data transformations and calculations.

Centralized data, automated ETL, created dashboards, and optimized production.

Vehicle Detection Using IBM Clusters:

  • Led a team of data scientists to set up a multi-GPU environment on IBM's cluster.

  • Configured and optimized NVIDIA GPUs for parallel processing to accelerate deep learning model training.

  • Developed and implemented convolutional neural networks (CNNs) using PyTorch and TensorFlow for vehicle detection from drone images.

  • Applied transfer learning techniques using pre-trained models such as YOLOv5 and Detectron2 to enhance detection accuracy and reduce training time.

  • Preprocessed and augmented large datasets of aerial images to improve model robustness and performance.

  • Utilized IBM Watson for additional cognitive computing capabilities and integration with cloud services.

Led team, developed CNNs for vehicle detection using IBM clusters.

Staff Mobility Tool

  • Created a comprehensive tool to optimize staff mobility in urban areas using advanced algorithms and graph theory.

  • Developed bio-inspired algorithms, as ant colony optimization, to find optimal routes and schedules for staff transportation.

  • Implemented the solution in Python, leveraging libraries such as NetworkX for graph-based analysis and Scikit-Learn for machine learning tasks.

  • Integrated Google Maps API for real-time geolocation data, route planning, and traffic updates to enhance the tool's accuracy and efficiency.

  • Reduced transportation costs by optimizing travel routes and schedules, resulting in significant savings for the company.

Colony bio-inspired algorithm optimizing AI

E-commerce Automation:

  • Developed and deployed automation scripts to manage and update the affiliate customer database, enhancing efficiency and reducing manual workload.

  • Utilized Apache Airflow to create DAGs (Directed Acyclic Graphs) for orchestrating complex data workflows, ensuring smooth and reliable execution of automated tasks.

  • Established secure FTPS connections for data transfer, ensuring the confidentiality and integrity of sensitive information.

  • Extracted and processed data from internal APIs, leveraging RESTful services to gather and update information dynamically.

  • Collaborated closely with backend development teams to integrate automation scripts seamlessly into the existing e-commerce platform.

Automated customer database updates, used Airflow, secure data transfer.

Instagram Automation Project:

  • Developed an automated system for managing an Instagram account, utilizing advanced AI models to generate and curate content.

  • Leveraged HuggingFace’s models for sentiment analysis , text generation and image generation ensuring relevant and engaging captions for posts.

Automated Instagram posts with AI for sentiment analysis and generation.

Video Animations Using Deforum and Stable Diffusion:

  • Created innovative video animations utilizing Deforum and Stable Diffusion for generating frame-by-frame image sequences.

  • Used Stable Diffusion models to produce high-quality, AI-generated images based on specific prompts and themes.

  • Integrated Deforum’s animation capabilities to interpolate and transition between frames smoothly, creating fluid and dynamic videos.

  • Fine-tuned open-source models for specific styles and subjects, enhancing the artistic quality and relevance of the animations.

Created animations with AI-generated images and smooth frame transitions.

Stock Market Forecasting:

  • eveloped a Jupyter notebook for forecasting stock market trends using machine learning techniques.

  • Leveraged TensorFlow to build and train predictive models based on historical stock data, focusing on time-series analysis.

  • Utilized LSTM (Long Short-Term Memory) networks to capture the temporal dependencies in stock prices, improving the accuracy of forecasts.

  • Integrated financial datasets from sources like Yahoo Finance and Alpha Vantage, ensuring comprehensive and up-to-date information for model training.

  • Visualized forecast results using Matplotlib and Seaborn, providing clear and interpretable charts for stakeholders.

Forecasted stock trends using LSTM and TensorFlow, visualized results.

Stock Market Trading Script:

  • Developed an automated trading script using Alpaca API, EC2, and Python to execute trades based on predefined strategies.

  • Implemented technical analysis strategies, including moving average crossovers, Bollinger Bands, and momentum indicators, to identify trading signals.

  • Utilized Alpaca’s commission-free trading API for real-time market data access and order execution.

  • Deployed the trading script on AWS EC2 to ensure high availability and performance, allowing for 24/7 operation.

  • Utilized Pandas and NumPy for data manipulation and analysis, ensuring efficient processing of large financial datasets.

Automated trading with Alpaca API and Python, implemented strategies.
bottom of page