Artificial Intelligence Fundamentals
Expert-defined terms from the Graduate Certificate in AI and GIS for Disaster Risk Reduction course at London School of Planning and Management. Free to read, free to share, paired with a globally recognised certification pathway.
Artificial Intelligence Fundamentals #
Artificial Intelligence Fundamentals
Artificial Intelligence (AI) Fundamentals refer to the basic concepts and princi… #
These fundamentals are essential for understanding how AI systems work and for developing applications that leverage AI capabilities. In the context of the Graduate Certificate in AI and GIS for Disaster Risk Reduction, understanding AI fundamentals is crucial for using AI techniques to analyze and mitigate disaster risks.
AI fundamentals include key concepts such as machine learning, deep learning, ne… #
These concepts are used to develop AI algorithms and models that can learn from data, make predictions, and solve complex problems.
Machine Learning #
Machine Learning
Machine Learning (ML) is a subset of AI that focuses on developing algorithms an… #
ML algorithms can identify patterns in data, make predictions based on new data, and continuously improve their performance over time.
ML algorithms can be categorized into supervised learning, unsupervised learning… #
Supervised learning algorithms learn from labeled data, unsupervised learning algorithms find patterns in unlabeled data, semi-supervised learning algorithms combine labeled and unlabeled data, and reinforcement learning algorithms learn through trial and error by receiving rewards for taking specific actions.
Deep Learning #
Deep Learning
Deep Learning is a subset of ML that uses artificial neural networks with multip… #
Deep Learning algorithms are capable of automatically learning hierarchical representations of data, which allows them to extract high-level features from raw input.
Deep Learning has been particularly successful in tasks such as image and speech… #
Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) are common architectures used in Deep Learning for handling image and sequential data, respectively.
Neural Networks #
Neural Networks
Neural Networks are computational models inspired by the structure and function… #
They consist of interconnected nodes organized in layers, where each node performs a simple computation and passes the result to nodes in the next layer. Neural Networks can learn complex patterns by adjusting the connections between nodes based on the input data.
Neural Networks are used in various AI applications, including image and speech… #
The process of training a Neural Network involves feeding it with labeled data, computing the prediction error, and updating the network parameters through backpropagation.
Natural Language Processing #
Natural Language Processing
Natural Language Processing (NLP) is a branch of AI that focuses on enabling com… #
NLP techniques are used to extract information from text, analyze sentiment, translate languages, and generate human-like responses in chatbots.
NLP applications include language translation services, sentiment analysis tools… #
NLP algorithms rely on techniques such as tokenization, part-of-speech tagging, named entity recognition, and syntactic parsing to process and analyze text data.
Computer Vision #
Computer Vision
Computer Vision is a field of AI that focuses on enabling computers to interpret… #
Computer Vision algorithms can analyze and process images and videos to recognize objects, detect patterns, and make decisions based on visual input.
Computer Vision is used in various applications, such as facial recognition syst… #
Convolutional Neural Networks (CNNs) are commonly used in Computer Vision for tasks like object detection, image classification, and image segmentation.
Reinforcement Learning #
Reinforcement Learning
Reinforcement Learning is a type of ML that involves training agents to make seq… #
Reinforcement Learning algorithms learn through trial and error by exploring different actions and learning which actions lead to the best outcomes.
Reinforcement Learning is used in applications such as game playing, robotics, a… #
The key components of Reinforcement Learning include the agent (decision-maker), environment (external system), actions (decisions made by the agent), rewards (feedback received by the agent), and policies (strategies for decision-making).
GIS #
GIS
Geographic Information System (GIS) is a technology that enables users to captur… #
GIS integrates geographical data with non-spatial data to provide insights into the spatial relationships and patterns within the data.
GIS is used in various fields, including urban planning, environmental managemen… #
GIS applications include mapping, spatial analysis, geocoding, routing, and spatial modeling. GIS software such as ArcGIS, QGIS, and Google Earth are commonly used for creating and analyzing spatial data.
Disaster Risk Reduction #
Disaster Risk Reduction
Disaster Risk Reduction (DRR) refers to efforts and strategies aimed at minimizi… #
DRR involves identifying, assessing, and reducing risks associated with natural hazards such as earthquakes, floods, hurricanes, and wildfires.
DRR activities include disaster preparedness, early warning systems, community r… #
GIS and AI technologies are increasingly being used in DRR to analyze and predict disaster risks, develop evacuation plans, and coordinate emergency response efforts.
Remote Sensing #
Remote Sensing
Remote Sensing is the process of acquiring information about the Earth's surface… #
Remote Sensing technologies capture images and data that can be used to monitor environmental changes, assess natural disasters, and support disaster response efforts.
Remote Sensing data is used in various applications, including land cover mappin… #
Remote Sensing techniques such as multispectral imaging, radar sensing, and LiDAR are used to collect different types of data for analyzing the Earth's surface.
Supervised Learning #
Supervised Learning
Supervised Learning is a type of ML where the algorithm learns from labeled trai… #
Supervised Learning algorithms aim to learn a mapping function that can predict the output labels for new, unseen data based on the input features.
Supervised Learning is used in tasks such as image classification, sentiment ana… #
Common supervised learning algorithms include Support Vector Machines (SVM), Random Forest, Logistic Regression, and Neural Networks.
Unsupervised Learning #
Unsupervised Learning
Unsupervised Learning is a type of ML where the algorithm learns from unlabeled… #
Unsupervised Learning algorithms aim to discover hidden patterns, structures, or relationships in the data without explicit guidance.
Unsupervised Learning is used in tasks such as clustering, anomaly detection, an… #
Common unsupervised learning algorithms include K-means clustering, Hierarchical clustering, Principal Component Analysis (PCA), and Autoencoders.
Semi #
Supervised Learning
Semi #
Supervised Learning is a type of ML that combines labeled and unlabeled data to train algorithms. Semi-Supervised Learning algorithms leverage the available labeled data along with a larger pool of unlabeled data to improve model performance and generalization.
Semi #
Supervised Learning is used in scenarios where obtaining labeled data is expensive or time-consuming. By using a mixture of labeled and unlabeled data, Semi-Supervised Learning algorithms can learn from both types of data and make better predictions on new, unseen data.
Overfitting #
Overfitting
Overfitting occurs when a machine learning model performs well on the training d… #
Overfitting happens when the model captures noise and irrelevant patterns in the training data, leading to poor performance on test data.
Overfitting can be mitigated by using techniques such as cross #
validation, regularization, and early stopping. Cross-validation helps evaluate the model's performance on multiple subsets of the data, regularization penalizes overly complex models, and early stopping stops training when the model starts overfitting.
Underfitting #
Underfitting
Underfitting occurs when a machine learning model is too simple to capture the u… #
Underfitting happens when the model is not complex enough to represent the relationships in the data.
Underfitting can be addressed by using more complex models, increasing the model… #
By increasing the model's complexity or adding more relevant features, underfitting can be reduced, and the model's performance can be improved.
Hyperparameter #
Hyperparameter
Hyperparameters are parameters that define the structure and behavior of a machi… #
Hyperparameters are set before training the model and are not learned from the data.
Hyperparameters need to be tuned to optimize the model's performance and general… #
Techniques such as grid search, random search, and Bayesian optimization can be used to search for the best hyperparameters for a given ML model. Hyperparameter tuning is essential for achieving the best results with machine learning algorithms.
Feature Engineering #
Feature Engineering
Feature Engineering is the process of selecting, transforming, and creating new… #
Feature Engineering involves selecting relevant features, encoding categorical variables, scaling numerical features, and creating new features through mathematical operations.
Feature Engineering plays a crucial role in the success of machine learning mode… #
Well-designed features can help the model capture the underlying patterns in the data, reduce overfitting, and improve prediction accuracy. Feature Engineering requires domain knowledge and creativity to extract meaningful information from the data.
Bias #
Variance Tradeoff
The Bias #
Variance Tradeoff is a key concept in machine learning that describes the balance between the model's bias (underfitting) and variance (overfitting). Models with high bias have limited capacity to capture the underlying patterns in the data, while models with high variance are sensitive to noise in the training data.
Finding the right balance between bias and variance is crucial for building mode… #
Techniques such as regularization, cross-validation, and ensemble learning can help manage the Bias-Variance Tradeoff and improve the model's performance on test data.
Ensemble Learning #
Ensemble Learning
Ensemble Learning is a machine learning technique that combines multiple base mo… #
Ensemble Learning methods create a diverse set of models and aggregate their predictions to make more robust and accurate predictions.
Ensemble Learning algorithms include Bagging, Boosting, and Stacking #
Bagging combines multiple models trained on different subsets of the data, Boosting trains models sequentially to correct errors made by previous models, and Stacking combines the predictions of multiple models using a meta-learner.
Deep Reinforcement Learning #
Deep Reinforcement Learning
Deep Reinforcement Learning is a combination of Deep Learning and Reinforcement… #
Deep Reinforcement Learning algorithms use deep neural networks to approximate value functions and policies for decision-making.
Deep Reinforcement Learning has been successful in applications such as playing… #
Deep Q-Networks (DQN), Proximal Policy Optimization (PPO), and Deep Deterministic Policy Gradient (DDPG) are popular Deep Reinforcement Learning algorithms.
Transfer Learning #
Transfer Learning
Transfer Learning is a machine learning technique that leverages knowledge from… #
Transfer Learning allows models to transfer learned features from one task to another, reducing the need for large amounts of labeled data.
Transfer Learning is used in scenarios where labeled data is scarce or when trai… #
By fine-tuning pre-trained models on new tasks or domains, Transfer Learning can accelerate the model training process and improve performance on new data.
Generative Adversarial Networks #
Generative Adversarial Networks
Generative Adversarial Networks (GANs) are a type of Deep Learning model that co… #
GANs learn to generate realistic samples by playing a game where the generator tries to create realistic samples, and the discriminator tries to distinguish between real and fake samples.
GANs are used in applications such as image generation, style transfer, and data… #
GANs have also been used to generate synthetic data for training machine learning models in scenarios where labeled data is limited or expensive to collect.
Recurrent Neural Networks #
Recurrent Neural Networks
Recurrent Neural Networks (RNNs) are a type of neural network architecture desig… #
RNNs have loops that allow information to persist and be passed from one time step to the next, enabling them to handle sequential data.
RNNs are used in tasks such as speech recognition, language modeling, and time s… #
Long Short-Term Memory (LSTM) and Gated Recurrent Units (GRU) are variants of RNNs that address the vanishing gradient problem and improve the model's ability to capture long-range dependencies in sequential data.
Convolutional Neural Networks #
Convolutional Neural Networks
Convolutional Neural Networks (CNNs) are a type of neural network architecture d… #
CNNs use convolutional layers to extract spatial features from the input data, pooling layers to reduce spatial dimensions, and fully connected layers for classification or regression tasks.
CNNs are used in tasks such as image classification, object detection, and image… #
CNN architectures like AlexNet, VGG, ResNet, and MobileNet have achieved state-of-the-art performance in computer vision tasks and have been widely adopted in various applications.
Autoencoders #
Autoencoders
Autoencoders are neural network architectures designed to learn efficient repres… #
Autoencoders consist of an encoder that maps the input data to a latent space representation and a decoder that reconstructs the input data from the latent space representation.
Autoencoders are used in tasks such as data compression, denoising, and feature… #
Variational Autoencoders (VAEs) and Generative Adversarial Networks (GANs) are extensions of traditional autoencoders that can generate new samples from learned representations.
Natural Language Generation #
Natural Language Generation
Natural Language Generation (NLG) is a subfield of NLP that focuses on generatin… #
NLG algorithms can convert data into coherent and grammatically correct sentences, paragraphs, or documents.
NLG is used in applications such as chatbots, language translation, and report g… #
NLG techniques include template-based generation, rule-based generation, and machine learning-based generation using models like recurrent neural networks and transformers.
Geospatial Analysis #
Geospatial Analysis
Geospatial Analysis is the process of analyzing and interpreting spatial data to… #
Geospatial Analysis combines GIS techniques with statistical, mathematical, and computational methods to extract insights from spatial data.
Geospatial Analysis is used in various fields, including urban planning, natural… #
Techniques such as spatial autocorrelation, hotspot analysis, interpolation, and network analysis are used to analyze spatial data and make informed decisions.
Spatial Data Visualization #
Spatial Data Visualization
Spatial Data Visualization is the process of creating visual representations of… #
Spatial Data Visualization techniques help users explore and interpret complex spatial relationships, patterns, and trends in geographical data.
Spatial Data Visualization tools include choropleth maps, heatmaps, scatter plot… #
GIS software like ArcGIS, QGIS, and Tableau are commonly used for creating interactive and informative visualizations of spatial data for decision-making and storytelling.
Geocoding #
Geocoding
Geocoding is the process of converting addresses or place names into geographic… #
Geocoding enables users to plot locations on maps, perform spatial queries, and analyze spatial relationships between different locations.
Geocoding is used in applications such as location #
based services, asset tracking, and emergency response. Geocoding services like Google Maps Geocoding API, OpenCage Geocoding API, and ArcGIS Geocoding Service provide accurate and reliable geocoding capabilities for various geospatial applications.
Spatial Analysis #
Spatial Analysis
Spatial Analysis is the process of examining spatial data to identify patterns,… #
Spatial Analysis involves applying statistical, mathematical, and computational methods to spatial data to derive meaningful insights and make informed decisions.
Spatial Analysis techniques include spatial autocorrelation, spatial interpolati… #
Spatial Analysis is used in various fields, including urban planning, environmental science, public health, and disaster management, to understand spatial patterns and relationships in the data.
Network Analysis #
Network Analysis
Network Analysis is the process of analyzing and visualizing relationships betwe… #
Network Analysis techniques help identify key nodes, clusters, and patterns in the network structure.
Network Analysis is used in applications such as social network analysis, supply… #
Network Analysis tools like Gephi, Cytoscape, and NetworkX enable users to analyze, visualize, and interpret complex network data for decision-making and strategic planning.
Spatial Autocorrelation #
Spatial Autocorrelation
Spatial Autocorrelation is a statistical measure that quantifies the degree of s… #
Spatial Autocorrelation helps identify spatial patterns, clusters, and outliers in geographical data.
Spatial Autocorrelation techniques include Global Moran's I, Local Moran's I, an… #
Global Moran's I measures the overall spatial autocorrelation in the data, Local Moran's I identifies local clusters of high or low values, and Geary's C detects spatial outliers and dispersion in the data.
Hotspot Analysis #
Hotspot Analysis
Hotspot Analysis is a spatial analysis technique that identifies clusters of hig… #
Hotspot Analysis helps detect areas with statistically significant patterns of concentration or dispersion, which can provide valuable insights for decision-making and planning.
Hotspot Analysis techniques include Getis #
Ord Gi* and Anselin Local Moran's I. Getis-Ord Gi* measures the degree of spatial clustering in the data, while Anselin Local Moran's I identifies local clusters of high or low values and their statistical significance.
Spatial Interpolation #
Spatial Interpolation
Spatial Interpolation is the process of estimating unknown values of a variable… #
Spatial Interpolation techniques help fill in missing data, create continuous surfaces, and generate maps of spatially distributed variables.
Spatial Interpolation techniques include Inverse Distance Weighting, Kriging, an… #
Inverse Distance Weighting estimates values based on the weighted average of neighboring points, Kriging uses a statistical model to estimate values with spatial dependence, and Splines create smooth surfaces based on local data points.
Geographic Information Science #
Geographic Information Science
Geographic Information Science (GIScience) is the scientific discipline that stu… #
GIScience integrates geography, cartography, computer science, and statistics to advance the understanding