keras-io

101 models ‱ 4 total models in database
Sort by:

monocular-depth-estimation

—
751
16

Deeplabv3p Resnet50

Multiclass semantic segmentation using DeepLabV3+ This repo contains the model and the notebook to this Keras example on Multiclass semantic segmentation using DeepLabV3+. The model is trained for demonstrative purposes and does not guarantee the best results in production. For better results, follow & optimize the Keras example as per your need. Background Information Semantic segmentation, with the goal to assign semantic labels to every pixel in an image, is an essential computer vision task. In this example, we implement the DeepLabV3+ model for multi-class semantic segmentation, a fully-convolutional architecture that performs well on semantic segmentation benchmarks. Training Data The model is trained on a subset (10,000 images) of Crowd Instance-level Human Parsing Dataset. The Crowd Instance-level Human Parsing (CIHP) dataset has 38,280 diverse human images. Each image in CIHP is labeled with pixel-wise annotations for 20 categories, as well as instance-level identification. This dataset can be used for the "human part segmentation" task. Model The model uses ResNet50 pretrained on ImageNet as the backbone model. References: 1. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation 2. Rethinking Atrous Convolution for Semantic Image Segmentation 3. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs

—
198
5

Ocr For Captcha

Keras Implementation of OCR model for reading captcha đŸ€–đŸŠčđŸ» This repo contains the model and the notebook to this Keras example on OCR model for reading captcha. Background Information This example demonstrates a simple OCR model built with the Functional API. Apart from combining CNN and RNN, it also illustrates how you can instantiate a new layer and use it as an "Endpoint layer" for implementing CTC loss. This model uses subclassing, learn more about subclassing from this guide.

—
117
81

lowlight-enhance-mirnet

—
93
35

sentiment-analysis

license:apache-2.0
54
1

Timeseries Anomaly Detection

This repo contains the model and the notebook to this Keras example on Timeseries anomaly detection using an Autoencoder. This script demonstrates how you can use a reconstruction convolutional autoencoder model to detect anomalies in timeseries data. We will use the Numenta Anomaly Benchmark(NAB) dataset. It provides artifical timeseries data containing labeled anomalous periods of behavior. Data are ordered, timestamped, single-valued metrics. The following hyperparameters were used during training: - optimizer: {'name': 'Adam', 'learningrate': 0.001, 'decay': 0.0, 'beta1': 0.9, 'beta2': 0.999, 'epsilon': 1e-07, 'amsgrad': False} - trainingprecision: float32 | Epochs | Train Loss | Validation Loss | |--- |--- |--- | | 1| 0.011| 0.014| | 2| 0.011| 0.015| | 3| 0.01| 0.012| | 4| 0.01| 0.013| | 5| 0.01| 0.012| | 6| 0.009| 0.014| | 7| 0.009| 0.013| | 8| 0.009| 0.012| | 9| 0.009| 0.012| | 10| 0.009| 0.011| | 11| 0.008| 0.01| | 12| 0.008| 0.011| | 13| 0.008| 0.009| | 14| 0.008| 0.011| | 15| 0.008| 0.009| | 16| 0.008| 0.009| | 17| 0.008| 0.009| | 18| 0.007| 0.01| | 19| 0.007| 0.009| | 20| 0.007| 0.008| | 21| 0.007| 0.009| | 22| 0.007| 0.008| | 23| 0.007| 0.008| | 24| 0.007| 0.007| | 25| 0.007| 0.008| | 26| 0.006| 0.009| | 27| 0.006| 0.008| | 28| 0.006| 0.009| | 29| 0.006| 0.008| ## Model Plot

—
51
71

Image-Classification-using-EANet

license:apache-2.0
39
1

CutMix_data_augmentation_for_image_classification

—
37
2

denoising-diffusion-implicit-models

—
30
10

structured-data-classification-grn-vsn

—
29
9

timeseries_forecasting_for_weather

—
27
21

Object-Detection-RetinaNet

—
26
19

video-classification-cnn-rnn

—
26
15

bert-semantic-similarity

—
26
9

PointNet

license:apache-2.0
25
5

low-light-image-enhancement

license:apache-2.0
23
86

drug-molecule-generation-with-VAE

—
22
12

TF_Decision_Trees

license:apache-2.0
19
6

timeseries-classification-from-scratch

—
19
3

vq-vae

—
18
2

ner-with-transformers

—
18
1

super-resolution

license:mit
14
32

imbalanced_classification

—
13
8

tab_transformer

—
12
41

video-vision-transformer

license:apache-2.0
12
7

vit_small_ds_v2

license:apache-2.0
12
1

cct

—
12
0

graph-attention-nets

—
11
6

GauGAN-Image-generation

—
11
4

collaborative-filtering-movielens

—
11
3

mobile-vit-xxs

—
11
0

text-generation-miniature-gpt

license:gpl
11
0

timeseries_transformer_classification

—
10
13

CycleGAN

—
10
9

dual-encoder-image-search

—
10
8

supervised-contrastive-learning-cifar10

license:apache-2.0
10
3

bidirectional-lstm-imdb

—
10
0

ppo-cartpole

—
10
0

vit-small-ds

license:apache-2.0
10
0

structured-data-classification

—
10
0

text-classification-with-transformer

—
10
0

consistency_training_with_supervision_teacher_model

—
10
0

wgan-molecular-graphs

—
9
5

deep-dream

—
9
3

video-transformers

—
9
2

VGG19

—
9
1

conv_autoencoder

license:gpl-3.0
9
0

time-series-anomaly-detection-autoencoder

—
8
15

cifar10_metric_learning

—
8
1

learning_to_tokenize_in_ViT

—
8
1

deep-deterministic-policy-gradient

—
8
0

semantic-segmentation

—
7
16

conv-lstm

—
7
4

attention_mil

—
7
1

keras-reptile

—
7
0

speaker-recognition

—
6
9

siamese-contrastive

—
6
5

Multimodal Entailment

Tensorflow Keras Implementation of Multimodal entailment. This repo contains the models Multimodal Entailment. In this example, we will build and train a model for predicting multimodal entailment. We will be using the multimodal entailment dataset recently introduced by Google Research. On social media platforms, to audit and moderate content we may want to find answers to the following questions in near real-time: Does a given piece of information contradict the other? Does a given piece of information imply the other? In NLP, this task is called analyzing textual entailment. However, that's only when the information comes from text content. In practice, it's often the case the information available comes not just from text content, but from a multimodal combination of text, images, audio, video, etc. Multimodal entailment is simply the extension of textual entailment to a variety of new input modalities.

—
6
4

MelGAN-spectrogram-inversion

—
6
2

ctc_asr

—
6
1

addition-lstm

—
6
1

SimSiam

—
6
1

consistency_training_with_supervision_student_model

—
6
0

conditional-gan

—
5
8

pointnet_segmentation

—
5
5

swin-transformers

—
5
4

semi-supervised-classification-simclr

license:apache-2.0
5
2

semantic-image-clustering

—
5
2

conv_mixer_image_classification

—
5
1

conv_Mixer

—
5
1

randaugment

license:apache-2.0
5
0

bit

—
5
0

transformers-qa

license:apache-2.0
4
4

english-speaker-accent-recognition-using-transfer-learning

—
4
2

pixel-cnn-mnist

—
4
1

Node2Vec_MovieLens

—
3
4

shiftvit

—
3
1

involution

license:mit
3
0

3D_CNN_Pneumonia

—
3
0

adamatch-domain-adaption

—
3
0

WGAN-GP

—
2
2

char-lstm-seq2seq

—
2
1

recommender-transformers

—
2
1

convmixer

license:apache-2.0
2
0

deit

—
2
0

cl_s2s

—
2
0

MPNN-for-molecular-property-prediction

—
1
3

nerf

—
1
1

dcgan

—
1
0

simple-mnist-convnet

—
1
0

gat-node-classification

—
1
0

real_nvp

—
1
0

transformer_asr

license:apache-2.0
1
0

image-captioning

—
0
8

EDSR

license:mit
0
7

ProbabalisticBayesianModel-Wine

—
0
2

neural-decision-forest

license:apache-2.0
0
2

dcgan-to-generate-face-images

—
0
2

pix2pix-generator

—
0
1

pix2pix-discriminator

—
0
1

neural_machine_translation_with_transformer

license:apache-2.0
0
1