Register
Login
Resources
Docs Blog Datasets Glossary Case Studies Tutorials & Webinars
Product
Data Engine LLMs Platform Enterprise
Pricing Explore
Connect to our Discord channel
Abigail Woolley d053484a7b
commented out github actions for CI/CD
3 weeks ago
d053484a7b
commented out github actions for CI/CD
3 weeks ago
78d8822db8
Model trainer added
4 weeks ago
45d82141f7
Model evaluation and mlflow added
3 weeks ago
45d82141f7
Model evaluation and mlflow added
3 weeks ago
src
b01b237f7c
CI/CD and web files added
3 weeks ago
b01b237f7c
CI/CD and web files added
3 weeks ago
b01b237f7c
CI/CD and web files added
3 weeks ago
24257c5c30
Updated file to ignore artifacts
1 month ago
b01b237f7c
CI/CD and web files added
3 weeks ago
96aecd3f88
Initial commit
1 month ago
5d3d426aea
updated file on how to run project's application
3 weeks ago
b01b237f7c
CI/CD and web files added
3 weeks ago
45d82141f7
Model evaluation and mlflow added
3 weeks ago
45d82141f7
Model evaluation and mlflow added
3 weeks ago
eac0e2bf88
Installed autopep8 for formatting
1 month ago
41d303af28
Added columns of the winequality dataset with corresponding data types
1 month ago
77e0a9eab3
reformatted files with autopep8
1 month ago
77e0a9eab3
reformatted files with autopep8
1 month ago
77e0a9eab3
reformatted files with autopep8
1 month ago
62d02ff83b
added winequality-data file
1 month ago
Storage Buckets

README.md

You have to be logged in to leave a comment. Sign In

End-to-end-Machine-Learning-Project-with-mlflow

Workflows

  1. Update config.yaml
  2. Update schema.yaml
  3. Update params.yaml
  4. Update the entity
  5. Update the configuration manager in src config
  6. Update the components
  7. Update the pipeline
  8. Update the main.py
  9. Update the app.py

How to run?

STEPS:

Clone the repository

https://github.com/ABIGAILDEBBY/End-to-end-Machine-Learning-Project-with-mlflow

STEP 01- Create a conda environment after opening the repository

conda create -n mlproj python=3.8 -y
conda activate mlproj

STEP 02- install the requirements

pip install -r requirements.txt
# Finally run the following command
python app.py

Now,

open up you local host and port

MLflow

Documentation

cmd
  • mlflow ui

dagshub

dagshub

MLFLOW_TRACKING_URI=https://dagshub.com/ABIGAILDEBBY/End-to-end-Machine-Learning-Project-with-mlflow.mlflow MLFLOW_TRACKING_USERNAME=ABIGAILDEBBY MLFLOW_TRACKING_PASSWORD=dbbeda6e65ada6d3eb722085d46d199f2af6de79 python script.py

Run this to export as env variables:


export MLFLOW_TRACKING_URI=https://dagshub.com/ABIGAILDEBBY/End-to-end-Machine-Learning-Project-with-mlflow.mlflow

export MLFLOW_TRACKING_USERNAME=ABIGAILDEBBY 

export MLFLOW_TRACKING_PASSWORD=dbbeda6e65ada6d3eb722085d46d199f2af6de79

AWS-CICD-Deployment-with-Github-Actions

1. Login to AWS console.

2. Create IAM user for deployment

#with specific access

1. EC2 access : It is virtual machine

2. ECR: Elastic Container registry to save your docker image in aws


#Description: About the deployment

1. Build docker image of the source code

2. Push your docker image to ECR

3. Launch Your EC2 

4. Pull Your image from ECR in EC2

5. Lauch your docker image in EC2

#Policy:

1. AmazonEC2ContainerRegistryFullAccess

2. AmazonEC2FullAccess

3. Create ECR repo to store/save docker image

- Save the URI: 566373416292.dkr.ecr.ap-south-1.amazonaws.com/mlproj

4. Create EC2 machine (Ubuntu)

5. Open EC2 and Install docker in EC2 Machine:

#optinal

sudo apt-get update -y

sudo apt-get upgrade

#required

curl -fsSL https://get.docker.com -o get-docker.sh

sudo sh get-docker.sh

sudo usermod -aG docker ubuntu

newgrp docker

6. Configure EC2 as self-hosted runner:

setting>actions>runner>new self hosted runner> choose os> then run command one by one

7. Setup github secrets:

AWS_ACCESS_KEY_ID=

AWS_SECRET_ACCESS_KEY=

AWS_REGION = us-east-1

AWS_ECR_LOGIN_URI = demo>>  566373416292.dkr.ecr.ap-south-1.amazonaws.com

ECR_REPOSITORY_NAME = simple-app

About MLflow

MLflow

  • Its Production Grade
  • Trace all of your expriements
  • Logging & tagging your model
Tip!

Press p or to see the previous file or, n or to see the next file

About

No description

Collaborators 1

Comments

Loading...