Skip to content

aretius/Machine_learning_Projects

Repository files navigation

Machine_learning_Projects

These are my Machine learning projects starting from the basics and moving further on to advanced level.

Project 1 :

Titanic Survival Exploration

• Created decision functions that attempt to predict survival outcomes from the 1912 Titanic disaster based on each passenger’s features, such as sex and age.
• Started with a simple algorithm and increased its complexity until I was able to accurately predict the outcomes for at least 80% of the passengers in the provided data.

Project 2 :

Predicting Boston Housing Prices

• Built a model to predict the value of a given house in the Boston real estate market using various statistical analysis tools.
• Identified the best price that a client can sell their house utilizing machine learning.

Project 3 :

Building a Student Intervention System

• Investigated the factors that affect a student's performance in high school.
• Trained and tested several supervised machine learning models on a given dataset to predict how likely a student is to pass.
• Selected the best model based on relative accuracy and efficiency.

Project 4 :

Creating Customer Segments

• Reviewed unstructured data to understand the patterns and natural categories that the data fits into.
• Used multiple algorithms and both empirically and theoretically compared and contrasted their results.
• Made predictions about the natural categories of multiple types in a dataset, then checked these predictions against the result of unsupervised analysis.

Project 5 :

Train a Smartcab How to Drive

• Applied reinforcement learning to build a simulated vehicle navigation agent.
• This project involved modeling a complex control problem in terms of limited available inputs, and designing a scheme to automatically learn an optimal driving strategy based on rewards and penalties.

Project 6 :

Digit Recognizer

• Kaggle Competitions -https://www.kaggle.com/c/digit-recognizer
• Task: Detect the written digits using data of various digits with labels from 0 - 10.
• Applied Random Forest with variable number of estimators and analyed it's relation with the accuracy.
• Another implementation involved using PCA and then KNN since KNN's are good for image realted problems.

Project 7 :

Ghouls, Goblins, and Ghosts... Boo!

• Kaggle Competition - https://www.kaggle.com/c/ghouls-goblins-and-ghosts-boo
• Task: to given features of creatures , classify them as Ghouls , Goblins & Ghosts.
• Used PCA & Logistic Regression with Hyper Parameter Optimization with an added SVM model , similar to stacking.

Project 8 :

Facial Keypoint Detection

• Kaggle Competition - https://www.kaggle.com/c/facial-keypoints-detection
• Task: to detect the facial keypoints such as nose_tip , eye_center etc given images with around 15 labelled keypoints as data.
• Applied CNN using librarires Lasagne & Theano and achieved 56th position on the leaderboard. • Used Dropout and compared various Optimizers and even used Data Generation technique for scale,size & rotation invariance.

Project 9:

House Prices: Advanced Regression Techniques

• Kaggle Competition - https://www.kaggle.com/c/house-prices-advanced-regression-techniques
• Task: With 79 explanatory variables describing (almost) every aspect of residential homes in Ames, Iowa, this competition challenges you to predict the final price of each home. • Used : Ensembling of models, XGboost, Data cleaning for 285 rank ( Other implementations are in the directory's README)

Project 10:

Two Sigma Connect: Rental Listing Inquiries

• Kaggle Competition - https://www.kaggle.com/c/two-sigma-connect-rental-listing-inquiries
• Task: predict how popular an apartment rental listing is based on the listing content like text description, photos, number of bedrooms, price, etc. The data comes from renthop.com, an apartment listing website.
• Used : - Xgboost, Hyper parameter optimization & Feature engineering ( since lots of data was text based , details in the notebook)

Project 11:

March Machine Learning Mania 2017

• Kaggle Competition - https://www.kaggle.com/c/march-machine-learning-mania-2017
Task:
• Stage 1 - You should submit predicted probabilities for every possible matchup in the past 4 NCAA tournaments (2013-2016).
• Stage 2 - You should submit predicted probabilities for every possible matchup before the 2017 tournament begins.

Project 12:

Two Sigma Financial Modeling Challenge

• Kaggle Competition - https://www.kaggle.com/c/two-sigma-financial-modeling
Task: • This dataset contains anonymized features pertaining to a time-varying value for a financial instrument.
• Each instrument has an id. Time is represented by the 'timestamp' feature and the variable to predict is 'y'.
• Used : - Selected best columns based upon Tree Algorithms, Removed Outliers by successive quantization on residuals and finally used Logistic Regression, Extra trees classifier & ensemble of linear models.

Project 13

• This Project is one of the Machine Learning Competitions on Hackerearth.
• Link : https://www.hackerearth.com/challenge/competitive/machine-learning-challenge-one/
• Predict : - The probability that a member will default of bank Indessa.
• Used : - Ensembling of LR ( With L1 Regularization ) and RF (300 estim) model with only selected features and simple Feature Engineering

Project 14

Indeed Machine Learning Codesprint hosted by Hackerrank
• Link : - https://www.hackerrank.com/indeed-ml-codesprint-2017
• Task: - Predicting Job tags with complete descriptions of jobs given.

Project 15

Sberbank Russian Housing Market

• Kaggle Competition : - https://www.kaggle.com/c/sberbank-russian-housing-market
• Task : - Identify the Realty price fluctuations in Russia's volatile economy
• Used : - Used Xgboost Model with hyper parameter optimization

About

Projects on Machine Learning

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published