Artifical Intelligence

Terminology

Neurons

Name Description
Vanilla Basic unit computing weighted sum + activation; used in FFNNs and CNNS
LSTM Advanced neuron with memory and gates for long-term dependencies in sequences

Layers

Name Description
Fully Connected Standard layer where each neuron connects to all inputs
Recurrent Maintains memory across timesteps; used in RNNs, LSTMs
Convolutional Extracts spatial features using filters; used in image data
Attention Computes weighted importance of different inputs; key in Transformers
Pooling Downsamples spatial data; used in CNNs to reduce size and noise
Normalization Stabilizes training by normalizing activations; includes BatchNorm, LayerNorm, GroupNorm
Dropout Randomly deactivates neurons during training to prevent overfitting

Activations

Name Description
ReLU max(0, x); fast, widely used in deep networks
Sigmoid S-shaped, output in (0, 1); used in binary classification
Tanh Like sigmoid but centered at 0; output in (-1, 1)
Softmax Outputs a probability distribution; used in final layer of multi-class classification

Networks

Name Description
Feedforward (FFNN) Basic architecture with no loops; used for static input-output tasks
RNN Handles sequential data using recurrence; remembers previous inputs
CNN Uses convolutions to process grid-like data such as images
Transformer Uses self-attention to model sequences without recurrence; state-of-the-art in NLP & beyond

Gradient Descent

Gradient Descent

Optimzier

Variations of gradient desents

Hyperparameters

Values that guide the training process

Reinforcement Learning

Updates parameters to maximize rewards

Variants of REINFORCE to improve stability and efficiency

Dataset

Machine Learning (ML)

ML on embedded devices

ML pipeline

Feature Selection and Extraction

Neural Networks and Training

Speech and Language Processing

Model Evaluation, Underfitting, and Overfitting

Anomaly Detection

Convolutional Neural Networks(CNN)

Sample Rate and Bit Depth

Mel Frequency Cepstral Coefficient(MFCC)

AI Accelerator

Image Classification and Neural Networks

CNN Visualizations and Data Augmentation

Transfer Learning

Object Detection

R-CNN

Advanced Image Processing

Neural network architecture example

import tensorflow as tf
from tensorflow.keras.models import Sequential
from tensorflow.keras.layers import Dense, InputLayer, Dropout, Conv1D, Conv2D, Flatten, Reshape, MaxPooling1D, MaxPooling2D, AveragePooling2D, BatchNormalization, Permute, ReLU, Softmax
from tensorflow.keras.optimizers.legacy import Adam

EPOCHS = args.epochs or 100
LEARNING_RATE = args.learning_rate or 0.005
# If True, non-deterministic functions (e.g. shuffling batches) are not used.
# This is False by default.
ENSURE_DETERMINISM = args.ensure_determinism
# this controls the batch size, or you can manipulate the tf.data.Dataset objects yourself
BATCH_SIZE = args.batch_size or 32
if not ENSURE_DETERMINISM:
    train_dataset = train_dataset.shuffle(buffer_size=BATCH_SIZE*4)
train_dataset=train_dataset.batch(BATCH_SIZE, drop_remainder=False)
validation_dataset = validation_dataset.batch(BATCH_SIZE, drop_remainder=False)

# model architecture
model = Sequential()
model.add(Reshape((int(input_length / 13), 13), input_shape=(input_length, )))
model.add(Conv1D(8, kernel_size=3, padding='same', activation='relu'))
model.add(MaxPooling1D(pool_size=2, strides=2, padding='same'))
model.add(Dropout(0.25))
model.add(Conv1D(16, kernel_size=3, padding='same', activation='relu'))
model.add(MaxPooling1D(pool_size=2, strides=2, padding='same'))
model.add(Dropout(0.25))
model.add(Flatten())
model.add(Dense(classes, name='y_pred', activation='softmax'))

# this controls the learning rate
opt = Adam(learning_rate=LEARNING_RATE, beta_1=0.9, beta_2=0.999)
callbacks.append(BatchLoggerCallback(BATCH_SIZE, train_sample_count, epochs=EPOCHS, ensure_determinism=ENSURE_DETERMINISM))

# train the neural network
model.compile(loss='categorical_crossentropy', optimizer=opt, metrics=['accuracy'])
model.fit(train_dataset, epochs=EPOCHS, validation_data=validation_dataset, verbose=2, callbacks=callbacks)

# Use this flag to disable per-channel quantization for a model.
# This can reduce RAM usage for convolutional models, but may have
# an impact on accuracy.
disable_per_channel_quantization = False

  1. big_data
  2. chatgpt
  3. data_analysis
  4. elasticsearch
  5. llm

Page Source