Python library for analysis of time series data including dimensionality reduction, clustering, and Markov model estimation

Overview

deeptime

License: LGPL v3 Build Status codecov

Releases:

Installation via conda recommended.

conda-forge PyPI
conda install -c conda-forge deeptime pip install deeptime

Documentation: deeptime-ml.github.io.

Building the latest trunk version of the package:

Using conda for dependency management and python setup.py:

git clone https://github.com/deeptime-ml/deeptime.git

cd deeptime
git submodule update --init

conda install numpy scipy cython scikit-learn pybind11

python setup.py install

Or using pip:

pip install git+https://github.com/deeptime-ml/[email protected]
Comments
  • VAMPnet partial fit

    VAMPnet partial fit

    I'd like to use VAMPnet for a large amount of data. I'm coming from pyEMMA where managing this is easy thanks to the function pyemma.coordinates.source. I see that deeptime is lacking this function but I do see the partial_fit function in almost all the functions. My problem is how this can be used in VAMPnet? The fit and partial_fit functions seem to do different things: in the first one for instance it is asked also for validation data while the second is satisfied by just the training data, same thing for the number of epochs. Another thing is whether I should fetch the model at the end. Right now I'm trying to do a loop over my data in the following way:

    import torch
    import torch.nn as nn
    import numpy as np
    from deeptime.data import TimeLaggedDataset
    from deeptime.util.torch import MLP
    from torch.utils.data import DataLoader
    from deeptime.decomposition.deep import VAMPNet
    
    lobe = MLP(units=ns, nonlinearity=nn.ReLU)
    vampnet = VAMPNet(lobe=lobe,learning_rate=3)
    # paths is just a list of strings containing the path to .npy data
    for path in paths:
        data = np.load(path)
        dataset = TimeLaggedDataset.from_trajectory(lagtime=500, data=data.astype(np.float32))
        lobe = MLP(units=ns, nonlinearity=nn.ReLU)
        vampnet = VAMPNet(lobe=lobe, learning_rate=1e-4)
        vampnet.partial_fit((dataset.data,dataset.data_lagged))
    model = vampnet.fetch_model()  # I'm pretty sure this is not right at all as most of the code before
    

    Note I'm not using the train_data and the val_data as you did in the documentation since partial_fit doesn't require it, but I'm pretty sure that I should somehow. I think that from the documentation is not clear how you should deal with this kind of problem. Thank you very much for your time

    opened by pl992 14
  • online covariance - example

    online covariance - example

    hi @brookehus !

    This looks like a very exciting library!

    Is there an example of the online covariance calulation.

    Its a problem I am looking at, at the moment.

    Kind regards, Andrew

    opened by andrewczgithub 14
  • Add VAMP, CKTests (MSM and VAMP)

    Add VAMP, CKTests (MSM and VAMP)

    This adds VAMP estimator/model and the infrastructure for lagged model validation (cktests).

    During the path of getting the stuff to work, I noticed that calling fit on an estimator has unexpected side effects. That is why we need to take a copy of it in LaggedModelValidator. The factory pattern however should make the need for this copy unnecessary, but because we encapsulate the current model instance, we can not work around this. @clonker do you think it would be sane to call _create_model upon fit() to avoid this kind of hassle? How would we enforce this behavior without interfering with overridden fit methods?

    opened by marscher 14
  • Installation runs into CUDA problem

    Installation runs into CUDA problem

    The bug reported at https://github.com/rusty1s/pytorch_sparse/issues/180 and https://github.com/pyg-team/pytorch_geometric/issues/4095 propagates to deeptime, too. Fixing PyTorch to older versions in the setup might help.

    opened by MQSchleich 13
  • Cannot install deeptime. I have Python 3.10 (Windows 10) and trying to install using Visual Studio 2022

    Cannot install deeptime. I have Python 3.10 (Windows 10) and trying to install using Visual Studio 2022

    Describe the bug A clear and concise description of what the bug is.

    Here's a quick checklist in what to include:

    • [x] Include a detailed description of the bug or suggestion
    • [x] pip list or conda list of the environment you are using (please attach a txt file to the issue).
    • [x] deeptime and operating system versions
    • [x] Minimal example if possible, a Python script, zipped input data (if not too large)
    opened by panandreou 12
  • Different VAMPNET results from CPU/GPU training

    Different VAMPNET results from CPU/GPU training

    Describe the bug I tried to run the ala2 notebook (https://github.com/deeptime-ml/deeptime-notebooks/blob/master/examples/ala2-example.ipynb) but ended up with quite different results with GPU v.s. CPU training. CPU had a much higher success rate and flat training curves compared to GPU. I am wondering if it is something common or if I had made any mistakes.

    Results I tested with 10 individual runs with the same parameters as in the tutorial notebook.

    • CPU cpu_training cpu_state

    • GPU gpu_training gpu_state

    System CPU: AMD EPYC 7551 GPU: RTX A5000 System: Ubuntu 20.04.1 Python 3.9 torch 1.11.0+cu113 deeptime '0.4.1+8.g38b0158.dirty' (main branch)

    opened by yuxuanzhuang 11
  • Sparse Identification of Nonlinear Dynamics

    Sparse Identification of Nonlinear Dynamics

    This pull request provides a simple implementation of the Sparse Identification of Nonlinear Dynamics (SINDy) method for nonlinear model discovery. It includes

    • A SINDy estimator class for fitting a SINDyModel objects to data
    • A SINDyModel class encapsulating the learned dynamical system
    • Tests for the important methods and use-cases

    The pull request is not quite complete—I still need to create an example showing how to use the new features—but I figured this was a good place for me to give you, the Scikit-time team, a chance to look at implementation. I tried to use existing sktime code as inspiration, but feel free to edit anything that doesn't conform to your standards or style.

    The other thing I'd like to flag is that I wasn't able to test whether the auto-generated documentation all looked okay. I welcome any suggestions on how to improve it.

    opened by briandesilva 11
  • Largest connected set returns a state when it should return an empty set

    Largest connected set returns a state when it should return an empty set

    The following should print the empty set [] but instead prints [3].

    import numpy as np
    from deeptime.markov import TransitionCountEstimator
    
    estimator = TransitionCountEstimator(lagtime=1, count_mode='sliding')
    
    full_counts_model = estimator.fit_fetch(np.asarray([0, 1, 2, 3]))
    submodel = full_counts_model.submodel_largest(
            connectivity_threshold=1,
            directed=True)
    print(submodel.state_symbols)
    
    opened by MaaikeG 10
  • Adding Koopman operator evaluation, ITS evaluation and CK test to VAMPnet module

    Adding Koopman operator evaluation, ITS evaluation and CK test to VAMPnet module

    I contacted you a few days ago regarding the VAMPnet usage. I solved everything but I think there's something missing. From the references it is reported that you could evaluate the ITS and the CKtest straight from the koopman matrix built on the transformed data from your VAMPnet. Starting from these data it isn't difficult to evaluate by your own the koopman operator, the ITS and the Chapmann-Kolmogorov test but for completeness I would ask these functions to be added so that the users could entrust your implementation, for sure much more robust and compact.

    Again congratulations for this huge and extremely well organized library!

    opened by pl992 10
  • TICA fit-once transform-many

    TICA fit-once transform-many

    Hi, in the past with pyEMMA I was able to fit once with TICA and project on different amount of dimensions after. Currently if I use

    tica = TICA(lagtime=20)
    tica.fit(mydata)
    tica.set_params(dim=3)
    tica.transform(mydata)
    

    after fitting it seems to ignore it and just returns the full dimensions (minus correlated features) Is there any way to achieve the same using deeptime? Since fitting takes lots of time it's quite useful to be able to test different number of dimensions without re-fitting the whole model.

    opened by stefdoerr 6
  • Documentation restructuring, some fixes, clustering docs

    Documentation restructuring, some fixes, clustering docs

    • The documentation is now split into apidocs and a more narrative documentation
    • the more narrative documentation is composed out of jupyter notebooks which are converted into html by the nbsphinx sphinx extension
    • added first draft of clustering narrative documentation
    • AMMs and OOMs are now part of the apidocs
    • in mini batch clustering: when calling fit() the previous behavior was falling back to ordinary k-means. Now it takes shuffled samples of the dataset and performs clustering on these mini batches
    opened by clonker 5
  • Allowing weights for VAMP dimensionality reduction

    Allowing weights for VAMP dimensionality reduction

    Is your feature request related to a problem? Please describe. The TICA and VAMP decomposition classes both provide similar interfaces for .fit_from_timeseries(data). However, the TICA class allows a weights argument.

    The VAMP decomposition, however, does not support weights, and throws an error if they're provided (see: https://github.com/deeptime-ml/deeptime/blob/11182accb1f8ce263f7c498b76c94bb657b5a998/deeptime/covariance/util/_running_moments.py#L245 )

    Describe the solution you'd like Support for weights in VAMP.

    I see some similarity between moments_XXXY() and moments_block(), but it seems like there was probably a reason for omitting support for weights from VAMP -- is that correct?

    opened by jdrusso 6
  • Sort Markov matrix

    Sort Markov matrix

    Is your feature request related to a problem? Please describe. I am doing Markov modelling for SAR/QSAR analysis of chemical compounds and would need sorted markov matrices.

    I suggest to sort the Markov matrix according to the most stable state. Something like with better memory management:

    def sort_markov_matrix(markov_matrix): 
        """Takes in random markov matrix
        returns sorted markov matrix 
    
        Args:
            markov_matrix (np.array): unsorted matrix
    
        Returns:
            (np.array): sorted Markov matrix
        """
        
        
        
        b = markov_matrix.copy()
        for i in range(len(markov_matrix)): 
            ref1 = markov_matrix[i,i]
            for j in range(i+1, len(markov_matrix)): 
                ref2 = markov_matrix[j, j]
                if ref2 > ref1: 
                    markov_matrix[i, :] = b[j, :]
                    markov_matrix[j, :] = b[i, :]
                    b = markov_matrix.copy()
                    for k in range(len(markov_matrix)):
                        markov_matrix[k,i] = b[k, j]
                        markov_matrix[k,j] = b[k, i]
                        b = markov_matrix.copy()
        return markov_matrix
    

    Test with

    
    def test_sort(): 
        a = np.array([[0.8, 0.1, 0.05, 0.05],[0.005, 0.9, 0.03, 0.015], [0.1, 0.2, 0.4, 0.3],[0.01, 0.02, 0.03, 0.94]])
        sorted_a = sort_markov_matrix(a)
        assert np.array_equal(sorted_a[0,:], np.array([0.94, 0.02, 0.01, 0.03])) == True, str(sorted_a[0,:])
        assert np.array_equal(sorted_a[1,:], np.array([0.015,0.9, 0.005, 0.03])) == True, str(sorted_a[1,:])
        assert np.array_equal(sorted_a[2,:], np.array([0.05, 0.1, 0.8, 0.05])) == True, str(sorted_a[2,:])
        assert np.array_equal(sorted_a[3,:], np.array([0.3, 0.2, 0.1, 0.4])) == True, str(sorted_a[3,:])
    
    

    What do you think?

    opened by MQSchleich 21
  • Feauture request - Tcorex covariance matrix

    Feauture request - Tcorex covariance matrix

    Hi All!

    love the libray so far.

    I was wondering if I could request a feature request of adding the calculation of the covariance matrix by this method below.

    https://github.com/hrayrhar/T-CorEx

    It does seem difficult in its current format and this api is so much easier to use.

    Kind regards, Andrew

    enhancement 
    opened by andrewczgithub 1
Releases(v0.4.4)
  • v0.4.4(Dec 20, 2022)

    What's Changed

    • remove print by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/257
    • Remove deprecated CK Test validator by @clonker in https://github.com/deeptime-ml/deeptime/pull/258
    • TV derivative and minor documentation fixes by @clonker in https://github.com/deeptime-ml/deeptime/pull/260
    • Lorenz system, Thomas attractor, Improvements regarding documentation, TV Derivative improvements by @clonker in https://github.com/deeptime-ml/deeptime/pull/261
    • Added an optional argument for specifying the tolerance in is_reversi… by @wehs7661 in https://github.com/deeptime-ml/deeptime/pull/262
    • Enable python 3.11 in CI pipelines by @clonker in https://github.com/deeptime-ml/deeptime/pull/263
    • pyproject.toml config by @clonker in https://github.com/deeptime-ml/deeptime/pull/264
    • Update for scikit-learn 1.2 by @clonker in https://github.com/deeptime-ml/deeptime/pull/267
    • cibuildwheel cleanup by @clonker in https://github.com/deeptime-ml/deeptime/pull/268

    New Contributors

    • @wehs7661 made their first contribution in https://github.com/deeptime-ml/deeptime/pull/262

    Full Changelog: https://github.com/deeptime-ml/deeptime/compare/v0.4.3...v0.4.4

    Source code(tar.gz)
    Source code(zip)
  • v0.4.3(Sep 8, 2022)

    What's Changed

    • CI: Update manifest to include npz data files, don't use editable install for nox make_docs by @clonker in https://github.com/deeptime-ml/deeptime/pull/233
    • plot energy surfaces by @clonker in https://github.com/deeptime-ml/deeptime/pull/234
    • Network plots by @clonker in https://github.com/deeptime-ml/deeptime/pull/236
    • plot fluxes by @clonker in https://github.com/deeptime-ml/deeptime/pull/237
    • Fix preprocess data call in SINDy by @clonker in https://github.com/deeptime-ml/deeptime/pull/240
    • Fix shape of empty count models by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/239
    • proper exception handling in trajectory-generating openmp block by @clonker in https://github.com/deeptime-ml/deeptime/pull/242
    • better error handling for trajectories dataset by @clonker in https://github.com/deeptime-ml/deeptime/pull/243
    • [ci] fix python 3.8 by @clonker in https://github.com/deeptime-ml/deeptime/pull/244
    • Convergence warning bugfix by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/245
    • Fix the bug by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/246
    • Contour and density plot utility functions by @clonker in https://github.com/deeptime-ml/deeptime/pull/247
    • py38 by @clonker in https://github.com/deeptime-ml/deeptime/pull/249
    • [setup] exclude examples, docs, devtools when packaging by @clonker in https://github.com/deeptime-ml/deeptime/pull/250
    • Var_cutoff can be disabled in covariance koopman models. Fixes #254. by @clonker in https://github.com/deeptime-ml/deeptime/pull/255

    Full Changelog: https://github.com/deeptime-ml/deeptime/compare/v0.4.2...v0.4.3

    Source code(tar.gz)
    Source code(zip)
  • v0.4.2(Apr 11, 2022)

    What's Changed

    • new tram notebook version by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/210
    • add device option for TAE class by @philipyoung9561 in https://github.com/deeptime-ml/deeptime/pull/211
    • Docs update by @clonker in https://github.com/deeptime-ml/deeptime/pull/213
    • Use CMake as primary build system via scikit-build by @clonker in https://github.com/deeptime-ml/deeptime/pull/215
    • Tram experiments by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/214
    • use Agg backend in plotting tests by @clonker in https://github.com/deeptime-ml/deeptime/pull/216
    • Fix bug in statdist caching for msms by @clonker in https://github.com/deeptime-ml/deeptime/pull/219
    • Tram experiments by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/217
    • typos in _koopman.py by @KirillShmilovich in https://github.com/deeptime-ml/deeptime/pull/221
    • Disable TF-32 tensor cores for VAMPNET training by @yuxuanzhuang in https://github.com/deeptime-ml/deeptime/pull/222
    • Small doc fixes and update testing dependencies by @clonker in https://github.com/deeptime-ml/deeptime/pull/223
    • pyproject.toml: use dynamic version by @clonker in https://github.com/deeptime-ml/deeptime/pull/226
    • Update versioneer and run cython from CMake by @clonker in https://github.com/deeptime-ml/deeptime/pull/228
    • [datasets] correct 1D triple well docs by @thempel in https://github.com/deeptime-ml/deeptime/pull/229
    • update project metadata by @clonker in https://github.com/deeptime-ml/deeptime/pull/230
    • Fix in setup by @clonker in https://github.com/deeptime-ml/deeptime/pull/231
    • update manifest to include cmakelists by @clonker in https://github.com/deeptime-ml/deeptime/pull/232

    New Contributors

    • @philipyoung9561 made their first contribution in https://github.com/deeptime-ml/deeptime/pull/211
    • @KirillShmilovich made their first contribution in https://github.com/deeptime-ml/deeptime/pull/221
    • @yuxuanzhuang made their first contribution in https://github.com/deeptime-ml/deeptime/pull/222

    Full Changelog: https://github.com/deeptime-ml/deeptime/compare/v0.4.1...v0.4.2

    Source code(tar.gz)
    Source code(zip)
  • v0.4.1(Feb 16, 2022)

    What's Changed

    • Update readme and use pyroject.toml in setup by @clonker in https://github.com/deeptime-ml/deeptime/pull/197
    • Hotfix tramdataset by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/196
    • Finish progress callbacks by @clonker in https://github.com/deeptime-ml/deeptime/pull/198
    • Log space sample weights by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/199
    • Implied timescales plot by @clonker in https://github.com/deeptime-ml/deeptime/pull/200
    • Make sample weights normalized by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/201
    • Fix numpy deprecated warning by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/204
    • New Chapman-Kolmogorov test implementation by @clonker in https://github.com/deeptime-ml/deeptime/pull/202
    • Update validation loss on TAE by @clarktemple03 in https://github.com/deeptime-ml/deeptime/pull/205
    • Fix the calculation of the PMF. by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/206
    • Mbar initialization for tram by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/207
    • Minor changes by @clonker in https://github.com/deeptime-ml/deeptime/pull/208
    • fix TRAM bug by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/209

    New Contributors

    • @clarktemple03 made their first contribution in https://github.com/deeptime-ml/deeptime/pull/205

    Full Changelog: https://github.com/deeptime-ml/deeptime/compare/v0.4.0...v0.4.1

    Source code(tar.gz)
    Source code(zip)
  • v0.4.0(Jan 24, 2022)

    What's Changed

    • Catch2 unit tests infrastructure by @clonker in https://github.com/deeptime-ml/deeptime/pull/170
    • Compile time index and include dir hierarchy by @clonker in https://github.com/deeptime-ml/deeptime/pull/171
    • Use mamba over conda for CI by @clonker in https://github.com/deeptime-ml/deeptime/pull/172
    • Fix mapping states back to symbols by @clonker in https://github.com/deeptime-ml/deeptime/pull/174
    • Minor docs updates by @clonker in https://github.com/deeptime-ml/deeptime/pull/176
    • Some cleanup refactoring by @clonker in https://github.com/deeptime-ml/deeptime/pull/177
    • Fix state assignments in HMM submodels by @thempel in https://github.com/deeptime-ml/deeptime/pull/178
    • Flaky test support by @clonker in https://github.com/deeptime-ml/deeptime/pull/179
    • TRAM by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/168
    • Minor documentation fixes for TRAM by @clonker in https://github.com/deeptime-ml/deeptime/pull/181
    • TOC and a bugfix by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/182
    • Unify header location and fix gil scoped release issues by @clonker in https://github.com/deeptime-ml/deeptime/pull/184
    • TRAM tests by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/183
    • Update installation instructions by @clonker in https://github.com/deeptime-ml/deeptime/pull/186
    • TRAM model by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/185
    • Bump numpy from 1.19.3 to 1.21.0 in /tests by @dependabot in https://github.com/deeptime-ml/deeptime/pull/187
    • Allow KMeans with no iterations by @thempel in https://github.com/deeptime-ml/deeptime/pull/189
    • TRAM docs and a unit test by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/188
    • TRAM docs - last little things by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/190
    • TRAM Progress bar handling by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/191
    • Fix submodel_disconnect and add test by @thempel in https://github.com/deeptime-ml/deeptime/pull/193
    • Refactor by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/192
    • Trajectory fragments: split on negative state indices by @MaaikeG in https://github.com/deeptime-ml/deeptime/pull/194
    • Cluster progress by @clonker in https://github.com/deeptime-ml/deeptime/pull/195

    New Contributors

    • @MaaikeG made their first contribution in https://github.com/deeptime-ml/deeptime/pull/168
    • @dependabot made their first contribution in https://github.com/deeptime-ml/deeptime/pull/187

    Full Changelog: https://github.com/deeptime-ml/deeptime/compare/v0.3.1...v0.4.0

    Source code(tar.gz)
    Source code(zip)
  • v0.3.1(Nov 15, 2021)

    Makes deeptime Python 3.10 ready

    What's Changed

    • Nox by @clonker in https://github.com/deeptime-ml/deeptime/pull/164
    • fix matmul import for old torch versions by @thempel in https://github.com/deeptime-ml/deeptime/pull/165
    • make docs with nox by @clonker in https://github.com/deeptime-ml/deeptime/pull/166
    • enable python 3.10 in build matrix by @clonker in https://github.com/deeptime-ml/deeptime/pull/167

    Full Changelog: https://github.com/deeptime-ml/deeptime/compare/v0.3.0...v0.3.1

    Source code(tar.gz)
    Source code(zip)
  • v0.3.0(Nov 1, 2021)

  • v0.2.3(Jan 29, 2021)

  • v0.2.1(Oct 26, 2020)

  • v0.2(Oct 23, 2020)

Robust Partial Matching for Person Search in the Wild

APNet for Person Search Introduction This is the code of Robust Partial Matching for Person Search in the Wild accepted in CVPR2020. The Align-to-Part

Yingji Zhong 36 Dec 18, 2022
Semi-supervised Semantic Segmentation with Directional Context-aware Consistency (CVPR 2021)

Semi-supervised Semantic Segmentation with Directional Context-aware Consistency (CAC) Xin Lai*, Zhuotao Tian*, Li Jiang, Shu Liu, Hengshuang Zhao, Li

Jia Research Lab 137 Dec 14, 2022
A short code in python, Enchpyter, is able to encrypt and decrypt words as you determine, of course

Enchpyter Enchpyter is a program do encrypt and decrypt any word you want (just letters). You enter how many letters jumps and write the word, so, the

João Assalim 2 Oct 10, 2022
Bulk2Space is a spatial deconvolution method based on deep learning frameworks

Bulk2Space Spatially resolved single-cell deconvolution of bulk transcriptomes using Bulk2Space Bulk2Space is a spatial deconvolution method based on

Dr. FAN, Xiaohui 60 Dec 27, 2022
A object detecting neural network powered by the yolo architecture and leveraging the PyTorch framework and associated libraries.

Yolo-Powered-Detector A object detecting neural network powered by the yolo architecture and leveraging the PyTorch framework and associated libraries

Luke Wilson 1 Dec 03, 2021
Official Python implementation of the 'Sparse deconvolution'-v0.3.0

Sparse deconvolution Python v0.3.0 Official Python implementation of the 'Sparse deconvolution', and the CPU (NumPy) and GPU (CuPy) calculation backen

Weisong Zhao 23 Dec 28, 2022
ICCV2021 Oral SA-ConvONet: Sign-Agnostic Optimization of Convolutional Occupancy Networks

Sign-Agnostic Convolutional Occupancy Networks Paper | Supplementary | Video | Teaser Video | Project Page This repository contains the implementation

64 Jan 05, 2023
A note taker for NVDA. Allows the user to create, edit, view, manage and export notes to different formats.

Quick Notetaker add-on for NVDA The Quick Notetaker add-on is a wonderful tool which allows writing notes quickly and easily anytime and from any app

5 Dec 06, 2022
State of the Art Neural Networks for Deep Learning

pyradox This python library helps you with implementing various state of the art neural networks in a totally customizable fashion using Tensorflow 2

Ritvik Rastogi 60 May 29, 2022
[AAAI 2022] Sparse Structure Learning via Graph Neural Networks for Inductive Document Classification

Sparse Structure Learning via Graph Neural Networks for inductive document classification Make graph dataset create co-occurrence graph for datasets.

16 Dec 22, 2022
A memory-efficient implementation of DenseNets

efficient_densenet_pytorch A PyTorch =1.0 implementation of DenseNets, optimized to save GPU memory. Recent updates Now works on PyTorch 1.0! It uses

Geoff Pleiss 1.4k Dec 25, 2022
An elaborate and exhaustive paper list for Named Entity Recognition (NER)

Named-Entity-Recognition-NER-Papers by Pengfei Liu, Jinlan Fu and other contributors. An elaborate and exhaustive paper list for Named Entity Recognit

Pengfei Liu 388 Dec 18, 2022
🚀 An end-to-end ML applications using PyTorch, W&B, FastAPI, Docker, Streamlit and Heroku

🚀 An end-to-end ML applications using PyTorch, W&B, FastAPI, Docker, Streamlit and Heroku

Made With ML 82 Jun 26, 2022
Putting NeRF on a Diet: Semantically Consistent Few-Shot View Synthesis Implementation

Putting NeRF on a Diet: Semantically Consistent Few-Shot View Synthesis Implementation This project attempted to implement the paper Putting NeRF on a

254 Dec 27, 2022
Leaderboard and Visualization for RLCard

RLCard Showdown This is the GUI support for the RLCard project and DouZero project. RLCard-Showdown provides evaluation and visualization tools to hel

Data Analytics Lab at Texas A&M University 246 Dec 26, 2022
PROJECT - Az Residential Real Estate Analysis

AZ RESIDENTIAL REAL ESTATE ANALYSIS -Decided on libraries to import. Includes pa

2 Jul 05, 2022
HyperaPy: An automatic hyperparameter optimization framework ⚡🚀

hyperpy HyperPy: An automatic hyperparameter optimization framework Description HyperPy: Library for automatic hyperparameter optimization. Build on t

Sergio Mora 7 Sep 06, 2022
Code of PVTv2 is released! PVTv2 largely improves PVTv1 and works better than Swin Transformer with ImageNet-1K pre-training.

Updates (2020/06/21) Code of PVTv2 is released! PVTv2 largely improves PVTv1 and works better than Swin Transformer with ImageNet-1K pre-training. Pyr

1.3k Jan 04, 2023
Pytorch Lightning 1.2k Jan 06, 2023
Introducing neural networks to predict stock prices

IntroNeuralNetworks in Python: A Template Project IntroNeuralNetworks is a project that introduces neural networks and illustrates an example of how o

Vivek Palaniappan 637 Jan 04, 2023