Simple and ready-to-use tutorials for TensorFlow

Overview

TensorFlow World

https://img.shields.io/badge/contributions-welcome-brightgreen.svg?style=flat https://badges.frapsoft.com/os/v2/open-source.svg?v=102 https://coveralls.io/repos/github/astorfi/TensorFlow-World/badge.svg?branch=master https://img.shields.io/twitter/follow/amirsinatorfi.svg?label=Follow&style=social

To support maintaining and upgrading this project, please kindly consider Sponsoring the project developer.

Any level of support is a great contribution here ❤️

This repository aims to provide simple and ready-to-use tutorials for TensorFlow. The explanations are present in the wiki associated with this repository.

Each tutorial includes source code and associated documentation.

Slack Group

Table of Contents

Motivation

There are different motivations for this open source project. TensorFlow (as we write this document) is one of / the best deep learning frameworks available. The question that should be asked is why has this repository been created when there are so many other tutorials about TensorFlow available on the web?

Why use TensorFlow?

Deep Learning is in very high interest these days - there's a crucial need for rapid and optimized implementations of the algorithms and architectures. TensorFlow is designed to facilitate this goal.

The strong advantage of TensorFlow is it flexibility in designing highly modular models which can also be a disadvantage for beginners since a lot of the pieces must be considered together when creating the model.

This issue has been facilitated as well by developing high-level APIs such as Keras and Slim which abstract a lot of the pieces used in designing machine learning algorithms.

The interesting thing about TensorFlow is that it can be found anywhere these days. Lots of the researchers and developers are using it and its community is growing at the speed of light! So many issues can be dealt with easily since they're usually the same issues that a lot of other people run into considering the large number of people involved in the TensorFlow community.

What's the point of this repository?

Developing open source projects for the sake of just developing something is not the reason behind this effort. Considering the large number of tutorials that are being added to this large community, this repository has been created to break the jump-in and jump-out process that usually happens to most of the open source projects, but why and how?

First of all, what's the point of putting effort into something that most of the people won't stop by and take a look? What's the point of creating something that does not help anyone in the developers and researchers community? Why spend time for something that can easily be forgotten? But how we try to do it? Even up to this very moment there are countless tutorials on TensorFlow whether on the model design or TensorFlow workflow.

Most of them are too complicated or suffer from a lack of documentation. There are only a few available tutorials which are concise and well-structured and provide enough insight for their specific implemented models.

The goal of this project is to help the community with structured tutorials and simple and optimized code implementations to provide better insight about how to use TensorFlow quick and effectively.

It is worth noting that, the main goal of this project is to provide well-documented tutorials and less-complicated code!

TensorFlow Installation and Setup the Environment

alternate text

In order to install TensorFlow please refer to the following link:

_img/mainpage/installation.gif

The virtual environment installation is recommended in order to prevent package conflict and having the capacity to customize the working environment.

TensorFlow Tutorials

The tutorials in this repository are partitioned into relevant categories.


Warm-up

alternate text

# topic Source Code  
1 Start-up Welcome / IPython Documentation

Basics

alternate text

# topic Source Code  
2 TensorFLow Basics Basic Math Operations / IPython Documentation
3 TensorFLow Basics TensorFlow Variables / IPython Documentation

Basic Machine Learning

alternate text

# topic Source Code  
4 Linear Models Linear Regression / IPython Documentation
5 Predictive Models Logistic Regression / IPython Documentation
6 Support Vector Machines Linear SVM / IPython  
7 Support Vector Machines MultiClass Kernel SVM / IPython  

Neural Networks

alternate text

# topic Source Code  
8 Multi Layer Perceptron Simple Multi Layer Perceptron / IPython  
9 Convolutional Neural Network Simple Convolutional Neural Networks Documentation
10 Autoencoder Undercomplete Autoencoder Documentation
11 Recurrent Neural Network RNN / IPython  

Some Useful Tutorials

Contributing

When contributing to this repository, please first discuss the change you wish to make via issue, email, or any other method with the owners of this repository before making a change. For typos, please do not create a pull request. Instead, declare them in issues or email the repository owner.

Please note we have a code of conduct, please follow it in all your interactions with the project.

Pull Request Process

Please consider the following criterions in order to help us in a better way:

  • The pull request is mainly expected to be a code script suggestion or improvement.
  • A pull request related to non-code-script sections is expected to make a significant difference in the documentation. Otherwise, it is expected to be announced in the issues section.
  • Ensure any install or build dependencies are removed before the end of the layer when doing a build and creating a pull request.
  • Add comments with details of changes to the interface, this includes new environment variables, exposed ports, useful file locations and container parameters.
  • You may merge the Pull Request in once you have the sign-off of at least one other developer, or if you do not have permission to do that, you may request the owner to merge it for you if you believe all checks are passed.

Final Note

We are looking forward to your kind feedback. Please help us to improve this open source project and make our work better. For contribution, please create a pull request and we will investigate it promptly. Once again, we appreciate your kind feedback and elaborate code inspections.

Acknowledgement

I have taken huge efforts in this project for hopefully being a small part of TensorFlow world. However, it would not have been plausible without the kind support and help of my friend and colleague Domenick Poster for his valuable advices. He helped me for having a better understanding of TensorFlow and my special appreciation goes to him.

Comments
  • TensorFlow

    TensorFlow

    $ git clone https://github.com/TensorFlow-World.git Cloning into 'TensorFlow-World'... remote: Not Found fatal: repository 'https://github.com/TensorFlow-World.git/' not found

    opened by ashu-22 13
  • RE: Policy regarding typos in codebase.

    RE: Policy regarding typos in codebase.

    This issue is regarding your policy regarding typos in your codebase. Here is the relevant section in your CONTRIBUTING.rst: For typos, please do not create a pull request. Instead, declare them in issues or email the repository owner.

    I suggest this policy be revised as it creates an extra step for you, the maintainer of this repo. For example, here is your current process:

    1. Contributor finds a typo.
    2. Contributor opens an issue.
    3. Repo owner reads the issue.
    4. Repo owner decides to create a code change to fix the typo and pushes the change.

    Here is the suggested process:

    1. Contributor finds a typo.
    2. Contributor creates a code change to fix the typo and creates a pull request
    3. Repo owner decides to accept the pull request and merges the changes.

    If typos can be discussed within a pull request, I don't see the point for a contributor to create an issue and then the repo owner creates a code change to fix the typo. I suggest using Github Issues to discuss lengthy proposals, but typos should be handled directly within a pull request. For example, see this Contributing guide for Github's open source guide.

    opened by adyavanapalli 4
  • Look for Python syntax errors or undefined names

    Look for Python syntax errors or undefined names

    • http://flake8.pycqa.org with find syntax errors and undefined names that can halt your program.
      • --select=E901,E999,F821,F822,F823 focuses the tool on the most critical issues
    • Fxxx codes are here: http://flake8.pycqa.org/en/latest/user/error-codes.html
    • Other codes are here: https://pycodestyle.readthedocs.io/en/latest/intro.html#error-codes
    • The output is here: https://travis-ci.org/astorfi/TensorFlow-World/builds/272817787

    F821 is really helpful for finding Python 2 / 3 differences but also for typos, copy/paste errors, etc.

    opened by cclauss 4
  • Update README.rst

    Update README.rst

    So, I cleaned up the grammar / spelling and got to the section about contributing to this repository.

    Based on this - it's definitely going to the top.

    Also. No. Here's your pull request.

    opened by razodactyl 3
  • logits is an undefined name in this context, should it be logits_last?

    logits is an undefined name in this context, should it be logits_last?

    Undefined names can raise NameErrorat runtime.

    https://travis-ci.org/astorfi/TensorFlow-World/jobs/272817788#L623-L626

    https://github.com/astorfi/TensorFlow-World/blob/master/codes/3-neural_networks/multi-layer-perceptron/code/test_classifier.py#L113

    opened by cclauss 2
  • train_op in linear regression

    train_op in linear regression

    Is defining train_op for each data point and epoch anew really needed? I'm new to TensorFlow so I can't tell why or why not this would make sense. For me, the regression seems to work fine (and much faster) if the line is removed.

    opened by mzur 2
  • sudo apt-get install nvidia-current-updates nvidia-settings-updates error

    sudo apt-get install nvidia-current-updates nvidia-settings-updates error

    Hello, just wanted to say this is a great guide. but when i execute : sudo apt-get install nvidia-current-updates nvidia-settings-updates its says: E: Unable to locate package nvidia-settings-updates

    can someone help me with this?

    opened by ghost 1
  • linear regression tutorial cost only reported for last data point

    linear regression tutorial cost only reported for last data point

    I noticed in the notebook for the linear regression that the cost was only being calculated for the last piece of data in each epoch.

    with tf.Session() as sess:
    
        # Initialize the variables[w and b].
        sess.run(tf.global_variables_initializer())
    
        # Get the input tensors
        X, Y = inputs()
    
        # Return the train loss and create the train_op.
        train_loss = loss(X, Y)
        train_op = train(train_loss)
    
        # Step 8: train the model
        for epoch_num in range(num_epochs): # run 100 epochs
            for x, y in data:
              train_op = train(train_loss)
    
              # Session runs train_op to minimize loss
              loss_value,_ = sess.run([train_loss,train_op], feed_dict={X: x, Y: y})
    
            # Displaying the loss per epoch.
            print('epoch %d, loss=%f' %(epoch_num+1, loss_value))
    
            # save the values of weight and bias
            wcoeff, bias = sess.run([W, b])
    

    data is being iterated over and the loss_value that is calculated is written over each time through the loop. Thus, the loss is only for the last piece of data. Since the loss needs to be computed over all of the data being used to train, the cost function should probably be something more like the following:

    def loss(X, Y):
        '''
        compute the loss by comparing the predicted value to the actual label.
        :param X: The inputs.
        :param Y: The labels.
        :return: The loss over the samples.
        '''
    
        # Making the prediction.
        Y_predicted = inference(X)
        return tf.reduce_sum(tf.squared_difference(Y, Y_predicted))/(2*data.shape[0])
    

    With this change above, the training section could be changed to the following (with the looping over data removed completely):

    with tf.Session() as sess:
    
        # Initialize the variables[w and b].
        sess.run(tf.global_variables_initializer())
    
        # Get the input tensors
        X, Y = inputs()
    
        # Return the train loss and create the train_op.
        train_loss = loss(X, Y)
        train_op = train(loss(X, Y))
    
        # Step 8: train the model
        for epoch_num in range(num_epochs): # run 100 epochs
            loss_value, _ = sess.run([train_loss,train_op], feed_dict={X: data[:,0], Y: data[:,1]})
    
            # Displaying the loss per epoch.
            print('epoch %d, loss=%f' %(epoch_num+1, loss_value))
    
            # save the values of weight and bias
            wcoeff, bias = sess.run([W, b])
    

    This would result in output like the following:

    epoch 1, loss=1573.599976
    epoch 2, loss=1332.513916
    epoch 3, loss=1128.868408
    epoch 4, loss=956.848999
    epoch 5, loss=811.544067
    

    I would be glad to submit a pull request with these and other minor changes. Please let me know if I have some misunderstanding.

    opened by mulhod 1
  • No Transformer Notebook

    No Transformer Notebook

    Hey,

    I see that there are no tutorial notebooks for Transformer implementations in this repository yet. Transformers are used primarily in the field of natural language processing. Like recurrent neural networks, Transformers are designed to handle sequential data, such as natural language, for tasks such as translation and text summarization.

    I would like to add such tutorial notebooks.

    opened by SauravMaheshkar 0
  • docs: fix simple typo, visualiaing -> visualising

    docs: fix simple typo, visualiaing -> visualising

    There is a small typo in docs/tutorials/1-basics/basic_math_operations/README.rst.

    Should read visualising rather than visualiaing.

    Semi-automated pull request generated by https://github.com/timgates42/meticulous/blob/master/docs/NOTE.md

    opened by timgates42 0
  • a small mistake in doc

    a small mistake in doc

    In the tutorial doc of chapter 1 "Basics/variables", there might be a misktake here:

    # "variable_list_custom" is the list of variables that we want to initialize.
    variable_list_custom = [weights, custom_variable]
    
    # The initializer
    init_custom_op = tf.variables_initializer(var_list=all_variables_list)
    

    The last line of the code above might end up with var_list=variable_list_custom, not all_variables_list.

    Here's url of the doc: https://github.com/astorfi/TensorFlow-World/tree/master/docs/tutorials/1-basics/variables#initializing-specific-variables Thank you for your repo, it helps me a lot.

    opened by Xiaokeai18 0
Releases(v1.0)
Owner
Amirsina Torfi
PhD & Developer working on Deep Learning, Computer Vision & NLP
Amirsina Torfi
A Python Reconnection Tool for alt:V

altv-reconnect What? It invokes a reconnect in the altV Client Dev Console. You get to determine when your local client should reconnect when developi

8 Jun 30, 2022
Computational modelling of ray propagation through optical elements using the principles of geometric optics (Ray Tracer)

Computational modelling of ray propagation through optical elements using the principles of geometric optics (Ray Tracer) Introduction By applying the

Son Gyo Jung 1 Jul 09, 2022
DALL-Eval: Probing the Reasoning Skills and Social Biases of Text-to-Image Generative Transformers

DALL-Eval: Probing the Reasoning Skills and Social Biases of Text-to-Image Generative Transformers Authors: Jaemin Cho, Abhay Zala, and Mohit Bansal (

Jaemin Cho 98 Dec 15, 2022
An attempt at the implementation of Glom, Geoffrey Hinton's new idea that integrates neural fields, predictive coding, top-down-bottom-up, and attention (consensus between columns)

GLOM - Pytorch (wip) An attempt at the implementation of Glom, Geoffrey Hinton's new idea that integrates neural fields, predictive coding,

Phil Wang 173 Dec 14, 2022
Image Lowpoly based on Centroid Voronoi Diagram via python-opencv and taichi

CVTLowpoly: Image Lowpoly via Centroid Voronoi Diagram Image Sharp Feature Extraction using Guide Filter's Local Linear Theory via opencv-python. The

Pupa 4 Jul 29, 2022
Configure SRX interfaces with Scrapli

Configure SRX interfaces with Scrapli Overview This example will show how to configure interfaces on Juniper's SRX firewalls. In addition to the Pytho

Calvin Remsburg 1 Jan 07, 2022
Pytorch implementation of paper: "NeurMiPs: Neural Mixture of Planar Experts for View Synthesis"

NeurMips: Neural Mixture of Planar Experts for View Synthesis This is the official repo for PyTorch implementation of paper "NeurMips: Neural Mixture

James Lin 101 Dec 13, 2022
Ensemble Visual-Inertial Odometry (EnVIO)

Ensemble Visual-Inertial Odometry (EnVIO) Authors : Jae Hyung Jung, Yeongkwon Choe, and Chan Gook Park 1. Overview This is a ROS package of Ensemble V

Jae Hyung Jung 95 Jan 03, 2023
Evaluating saliency methods on artificial data with different background types

Evaluating saliency methods on artificial data with different background types This repository contains the relevant code for the MedNeurips 2021 subm

2 Jul 05, 2022
Implementation of "Unsupervised Domain Adaptive 3D Detection with Multi-Level Consistency"

Unsupervised Domain Adaptive 3D Detection with Multi-Level Consistency (ICCV2021) Paper Link: https://arxiv.org/abs/2107.11355 This implementation bui

32 Nov 17, 2022
Architecture Patterns with Python (TDD, DDD, EDM)

architecture-traning Architecture Patterns with Python (TDD, DDD, EDM) Chapter 5. 높은 기어비와 낮은 기어비의 TDD 5.2 도메인 계층 테스트를 서비스 계층으로 옮겨야 하는가? 도메인 계층 테스트 def

minsung sim 2 Mar 04, 2022
Codebase for the paper titled "Continual learning with local module selection"

This repository contains the codebase for the paper Continual Learning via Local Module Composition. Setting up the environemnt Create a new conda env

Oleksiy Ostapenko 20 Dec 10, 2022
Pytorch Implementation of Google's Parallel Tacotron 2: A Non-Autoregressive Neural TTS Model with Differentiable Duration Modeling

Parallel Tacotron2 Pytorch Implementation of Google's Parallel Tacotron 2: A Non-Autoregressive Neural TTS Model with Differentiable Duration Modeling

Keon Lee 170 Dec 27, 2022
This is a repository for a semantic segmentation inference API using the OpenVINO toolkit

BMW-IntelOpenVINO-Segmentation-Inference-API This is a repository for a semantic segmentation inference API using the OpenVINO toolkit. It's supported

BMW TechOffice MUNICH 34 Nov 24, 2022
Blender Add-On for slicing meshes with planes

MeshSlicer Blender Add-On for slicing meshes with multiple overlapping planes at once. This is a simple Blender addon to slice a silmple mesh with mul

52 Dec 12, 2022
A Python package for time series augmentation

tsaug tsaug is a Python package for time series augmentation. It offers a set of augmentation methods for time series, as well as a simple API to conn

Arundo Analytics 278 Jan 01, 2023
Code for the paper "Functional Regularization for Reinforcement Learning via Learned Fourier Features"

Reinforcement Learning with Learned Fourier Features State-space Soft Actor-Critic Experiments Move to the state-SAC-LFF repository. cd state-SAC-LFF

Alex Li 10 Nov 11, 2022
Some simple programs built in Python: webcam with cv2 that detects eyes and face, with grayscale filter

Programas en Python Algunos programas simples creados en Python: 📹 Webcam con c

Madirex 1 Feb 15, 2022
Accommodating supervised learning algorithms for the historical prices of the world's favorite cryptocurrency and boosting it through LightGBM.

Accommodating supervised learning algorithms for the historical prices of the world's favorite cryptocurrency and boosting it through LightGBM.

1 Nov 27, 2021
VLG-Net: Video-Language Graph Matching Networks for Video Grounding

VLG-Net: Video-Language Graph Matching Networks for Video Grounding Introduction Official repository for VLG-Net: Video-Language Graph Matching Networ

Mattia Soldan 25 Dec 04, 2022