Python Neural Networks diabetes

USE Jupyter Notebook
Given Code: ‘Part of answer’ but incomplete…..
# Importing libraries

import pandas as pd

import numpy as np

from sklearn.neighbors import KNeighborsClassifier

from sklearn.tree import DecisionTreeClassifier

from sklearn.model_selection import train_test_split

from sklearn import metrics

# Loading the dataset

data = pd.read_csv(‘diabetes.csv’)

# Separating features and target

features = data[[‘Pregnancies’, ‘Glucose’, ‘BloodPressure’, ‘SkinThickness’, ‘Insulin’, ‘BMI’, ‘DiabetesPedigreeFunction’, ‘Age’]]

target = data[‘Outcome’]

# Splitting training and testing data in the ratio 70 : 30 respectively

x_train, x_test, y_train, y_test = train_test_split(features, target, random_state=0, test_size=0.3)

# KNN model

modelKNN = KNeighborsClassifier()

# Train the model, y_train)

# Predict test set

y_pred_knn = modelKNN.predict(x_test)

# Print accuracy of the KNN model


# Print Confusion matrix

print(metrics.confusion_matrix(y_test, y_pred_knn))

# Print Classification Report for determining precision, recall, f1-score and support of the model

print(metrics.classification_report(y_test, y_pred_knn))

# Decision Tree model

modelDT = DecisionTreeClassifier()

# Train the model, y_train)

# Predict test set

y_pred_dt = modelDT.predict(x_test)

# Print accuracy of the Decision Tree model


# Print Confusion matrix

print(metrics.confusion_matrix(y_test, y_pred_dt))

# Print Classification Report for determining precision, recall, f1-score and support of the model

print(metrics.classification_report(y_test, y_pred_dt))
Please do the data cleaning and feature selection…
Please use the given code/layout given below original question….

“”for more info check original .ipynb file””
Fill out the code below so that it creates a multi-layer perceptron with a single hidden layer (with 4 nodes) and trains it via back-propagation. Specifically your code should: 

Initialize the weights to random values between -1 and 1
Perform the feed-forward computation
Compute the loss function
Calculate the gradients for all the weights via back-propagation
Update the weight matrices (using a learning_rate parameter)
Execute steps 2-5 for a fixed number of iterations
Plot the accuracies and log loss and observe how they change over time

First, initialise the parameters. This means determining the following:
Size of the network
Hidden layers (or a single hidden layer in this example) can be any size. Layers with more neurons are more powerful, but also more likely to overfit, and take longer to train. The output layer size corresponds to the number of classes.
Number of iterations
This parameter determines how many times the network will be updated.
Learning rate
Each time we update the weights, we do so by taking a step into the direction that we calculated will improve the accuracy of the network. The size of that step is determined by the learning rate. Taking small steps will slow the process down, but taking steps that are too large can cause results to vary wildly and not reach a stable optimum.
Next, fill in the code below to train a multi-layer perceptron and see if it correctly classies the input.
in[]# Reshape y

y =
# Initializing weights
W_1 =

W_2 =
# Definining number of iterations and learning rate(‘lr’)

num_iter =
learning rate =
in[]# Creating empty lists for loss values(error) and accuracy

loss_vals, accuracies = [], []
for j in range(num_iter):

# Do a forward pass through the dataset and compute the loss

# Decide on intervals and add on the current loss and accuracy to the respective list
# Update the weights
in[]# Plot the loss values and accuracy

Higher number of iterations and lower learning rate seems to improve accuracy

Place your order
(550 words)

Approximate price: $22

Calculate the price of your order

550 words
We'll send you the first draft for approval by September 11, 2018 at 10:52 AM
Total price:
The price is based on these factors:
Academic level
Number of pages
Basic features
  • Free title page and bibliography
  • Unlimited revisions
  • Plagiarism-free guarantee
  • Money-back guarantee
  • 24/7 support
On-demand options
  • Writer’s samples
  • Part-by-part delivery
  • Overnight delivery
  • Copies of used sources
  • Expert Proofreading
Paper format
  • 275 words per page
  • 12 pt Arial/Times New Roman
  • Double line spacing
  • Any citation style (APA, MLA, Chicago/Turabian, Harvard)

Our guarantees

Delivering a high-quality product at a reasonable price is not enough anymore.
That’s why we have developed 5 beneficial guarantees that will make your experience with our service enjoyable, easy, and safe.

Money-back guarantee

You have to be 100% sure of the quality of your product to give a money-back guarantee. This describes us perfectly. Make sure that this guarantee is totally transparent.

Read more

Zero-plagiarism guarantee

Each paper is composed from scratch, according to your instructions. It is then checked by our plagiarism-detection software. There is no gap where plagiarism could squeeze in.

Read more

Free-revision policy

Thanks to our free revisions, there is no way for you to be unsatisfied. We will work on your paper until you are completely happy with the result.

Read more

Privacy policy

Your email is safe, as we store it according to international data protection rules. Your bank details are secure, as we use only reliable payment systems.

Read more

Fair-cooperation guarantee

By sending us your money, you buy the service we provide. Check out our terms and conditions if you prefer business talks to be laid out in official language.

Read more