Introduction to Machine Learning

Expert-defined terms from the Graduate Certificate in Machine Learning in Polymer Science and Engineering course at London School of Planning and Management. Free to read, free to share, paired with a globally recognised certification pathway.

Introduction to Machine Learning

Introduction to Machine Learning in Polymer Science and Engineering #

Introduction to Machine Learning in Polymer Science and Engineering

In the Graduate Certificate in Machine Learning in Polymer Science and Engineeri… #

Machine learning is a subfield of artificial intelligence that focuses on developing algorithms and statistical models that enable computers to learn from and make predictions or decisions based on data without being explicitly programmed. In the context of polymer science and engineering, machine learning can be used to analyze complex data sets, predict material properties, optimize processes, and design new materials with desired characteristics.

Adversarial Learning #

Adversarial Learning

Adversarial learning is a machine learning technique where two neural networks,… #

The generator generates synthetic data samples, while the discriminator tries to distinguish between real and fake samples. This technique is commonly used in generating realistic images, improving the performance of models, and enhancing data security.

Backpropagation #

Backpropagation

Backpropagation is a fundamental algorithm used to train neural networks by adju… #

It involves calculating the gradient of the loss function with respect to the weights and updating the weights in the opposite direction of the gradient to minimize the error. Backpropagation is essential for optimizing neural networks and improving their predictive accuracy.

Clustering #

Clustering

Clustering is a machine learning technique used to group similar data points tog… #

It is an unsupervised learning method that aims to identify patterns or structures in the data without the need for labeled output. Clustering algorithms such as K-means, hierarchical clustering, and DBSCAN are commonly used in polymer science and engineering to discover relationships among materials, classify polymers, and segment data for analysis.

Dimensionality Reduction #

Dimensionality Reduction

Dimensionality reduction is the process of reducing the number of features or va… #

This technique is used to overcome the curse of dimensionality, improve computational efficiency, and visualize high-dimensional data in lower dimensions. Principal component analysis (PCA), t-distributed stochastic neighbor embedding (t-SNE), and autoencoders are popular methods for dimensionality reduction in machine learning.

Ensemble Learning #

Ensemble Learning

Ensemble learning is a machine learning technique that combines multiple models,… #

By aggregating the predictions of individual models through voting or averaging, ensemble methods can reduce overfitting, increase robustness, and enhance accuracy. Common ensemble techniques include random forests, boosting, and bagging, which are widely used in polymer science and engineering for predictive modeling and classification tasks.

Feature Engineering #

Feature Engineering

Feature engineering is the process of selecting, transforming, and creating new… #

It involves identifying relevant features, handling missing values, encoding categorical variables, scaling numerical attributes, and generating interaction terms. Effective feature engineering can lead to better model interpretability, generalization, and predictive power in polymer science and engineering applications.

Gradient Descent #

Gradient Descent

Gradient descent is an optimization algorithm used to minimize the loss function… #

It calculates the gradient of the loss function with respect to the model parameters and adjusts the weights in the direction of the steepest descent to find the global or local minimum. Gradient descent variants such as stochastic gradient descent (SGD), mini-batch gradient descent, and Adam optimization are commonly employed in training deep learning models.

Hyperparameter Tuning #

Hyperparameter Tuning

Hyperparameter tuning is the process of optimizing the hyperparameters of a mach… #

Hyperparameters are configuration settings that are not learned during training, such as learning rate, batch size, regularization strength, and network architecture. Techniques like grid search, random search, and Bayesian optimization are used to search the hyperparameter space efficiently and fine-tune models in polymer science and engineering.

Imbalanced Data #

Imbalanced Data

Imbalanced data refers to a situation where the distribution of classes in a dat… #

This imbalance can lead to biased models, poor predictive performance, and difficulty in detecting minority classes. Techniques like oversampling, undersampling, SMOTE (Synthetic Minority Over-sampling Technique), and class weighting are used to address imbalanced data challenges in classification tasks in polymer science and engineering.

Joint Distribution #

Joint Distribution

Joint distribution is a probability distribution that describes the likelihood o… #

It specifies the probability of all possible combinations of values for the variables in the system. Understanding the joint distribution of variables is essential for modeling dependencies, calculating conditional probabilities, and making predictions in machine learning applications. In polymer science and engineering, joint distributions can help analyze the relationships between material properties and processing conditions.

K #

nearest Neighbors (KNN)

K-nearest neighbors (KNN) is a simple yet effective machine learning algorithm u… #

It assigns a new data point to the class that is most common among its k nearest neighbors in feature space. KNN is a non-parametric method that does not require training a model and can handle complex decision boundaries. In polymer science and engineering, KNN is used for clustering, pattern recognition, and similarity-based analysis of materials.

Loss Function #

Loss Function

A loss function is a mathematical function that measures the error or discrepanc… #

It quantifies how well the model is performing during training and guides the optimization process by providing feedback on the model's performance. Common loss functions include mean squared error (MSE), cross-entropy loss, hinge loss, and KL divergence, which are tailored to specific tasks such as regression, classification, and generative modeling in polymer science and engineering.

Model Selection #

Model Selection

Model selection is the process of choosing the best machine learning model from… #

It involves evaluating models on training, validation, and test datasets to ensure they can make accurate predictions on unseen data. Techniques like cross-validation, grid search, and Bayesian model averaging are used to compare models, prevent overfitting, and select the most appropriate model for polymer science and engineering applications.

Neural Network #

Neural Network

A neural network is a computational model inspired by the structure and function… #

Neural networks are capable of learning complex patterns and relationships in data through the process of training with labeled examples. Deep neural networks, such as convolutional neural networks (CNNs) and recurrent neural networks (RNNs), have revolutionized machine learning and are widely used in polymer science and engineering for image analysis, sequence prediction, and material design.

Overfitting #

Overfitting

Overfitting occurs when a machine learning model performs well on training data… #

It is caused by the model fitting noise or irrelevant patterns in the training set, leading to poor performance on test data. Regularization techniques, early stopping, dropout, and cross-validation are employed to prevent overfitting, improve model robustness, and ensure reliable predictions in polymer science and engineering applications.

Principle of Occam's Razor #

Principle of Occam's Razor

Occam's razor is a principle in philosophy and science that states that among co… #

In the context of machine learning, Occam's razor suggests that simpler models are more likely to generalize well and avoid overfitting compared to complex models. By prioritizing simplicity and interpretability, machine learning practitioners can build more effective models and make better decisions in polymer science and engineering.

Quantum Machine Learning #

Quantum Machine Learning

Quantum machine learning is an emerging interdisciplinary field that combines qu… #

Quantum computers leverage the principles of quantum mechanics to perform computations on quantum bits (qubits) and explore vast solution spaces in parallel. Quantum machine learning algorithms, such as quantum support vector machines and quantum neural networks, hold promise for accelerating material discovery, optimization, and simulation in polymer science and engineering.

Reinforcement Learning #

Reinforcement Learning

Reinforcement learning is a branch of machine learning that focuses on training… #

It is based on the concept of trial and error learning, where agents learn optimal policies through exploration and exploitation. Reinforcement learning algorithms like Q-learning, deep Q-networks (DQN), and policy gradients are used in polymer science and engineering for process optimization, adaptive control, and autonomous material design.

Supervised Learning #

Supervised Learning

Supervised learning is a machine learning paradigm where the model is trained on… #

The goal is to learn a mapping function from input features to output labels to make predictions on unseen data accurately. Supervised learning tasks include regression, classification, and ranking, where the model learns from examples provided by a teacher or supervisor. In polymer science and engineering, supervised learning is applied to predict material properties, classify polymers, and optimize manufacturing processes.

Transfer Learning #

Transfer Learning

Transfer learning is a machine learning technique that leverages knowledge learn… #

By transferring features, representations, or parameters from a pre-trained model to a new model, transfer learning can expedite training, reduce data requirements, and enhance generalization. In polymer science and engineering, transfer learning is used to transfer knowledge from materials databases, domain-specific models, or related fields to accelerate material discovery and property prediction.

Unsupervised Learning #

Unsupervised Learning

Unsupervised learning is a machine learning approach where the model learns patt… #

The goal is to discover intrinsic patterns, clusters, or anomalies in the data and extract meaningful insights. Unsupervised learning techniques like clustering, dimensionality reduction, and generative modeling are valuable for exploring uncharted territories, identifying hidden trends, and uncovering novel relationships in polymer science and engineering datasets.

Variational Inference #

Variational Inference

Variational inference is a probabilistic method used to approximate complex post… #

It involves formulating an approximate distribution that minimizes the Kullback-Leibler (KL) divergence with the true posterior distribution. Variational inference is employed in Bayesian neural networks, latent variable models, and hierarchical Bayesian models to estimate uncertain or latent variables, perform model selection, and quantify uncertainty in predictions. In polymer science and engineering, variational inference can be used to model material properties, predict performance under uncertainty, and optimize experimental designs.

Word Embedding #

Word Embedding

Word embedding is a technique in natural language processing (NLP) that represen… #

Word embeddings capture semantic relationships and contextual information between words, enabling machines to understand and process textual data effectively. Popular word embedding models such as Word2Vec, GloVe, and fastText are used in sentiment analysis, language modeling, and information retrieval tasks in polymer science and engineering to analyze research articles, patents, and technical documents.

XGBoost #

XGBoost

XGBoost is an optimized implementation of gradient boosting, a machine learning… #

XGBoost employs a scalable and efficient tree boosting technique that minimizes loss functions and regularizes models to prevent overfitting. It is widely used in classification, regression, and ranking tasks in polymer science and engineering due to its speed, accuracy, and interpretability.

May 2026 cohort · 29 days left
from £99 GBP
Enrol