Neural networks are computing systems inspired by the human brain’s biological neural networks. They process vast amounts of data, identifying patterns and learning from them, much like the human brain learns from experience. This ability to learn and adapt makes neural networks a crucial component in the field of artificial intelligence (AI), powering advancements in areas like speech recognition, image processing, and autonomous vehicles. Understanding neural networks requires a deeper look into AI and its subfield, deep learning.
In the ever-evolving landscape of technology, one term that has been gaining significant traction is ‘Neural Network’. But what exactly is a neural network? How does it function? And why is it so crucial in the realm of artificial intelligence (AI)? This article aims to demystify the concept of neural networks, providing a comprehensive understanding of this complex yet fascinating subject.
Neural networks, in essence, are computing systems inspired by the human brain’s biological neural networks. They are designed to ‘learn’ by processing vast amounts of data and identifying underlying patterns, much like how our brain learns from experience. This ability to learn and adapt makes neural networks a vital component in the field of AI, powering advancements in areas like speech recognition, image processing, and even autonomous vehicles.
However, to fully appreciate the significance of neural networks, it’s essential to delve deeper into the broader context of AI and its subfield, deep learning. AI, a concept that has been around since the mid-20th century, is the science of making machines mimic human intelligence. Deep learning, on the other hand, is a more recent development. It’s a subset of AI that focuses on training neural networks to learn from unstructured and unlabeled data, thereby enabling machines to make decisions without being explicitly programmed.
The origins of AI and the evolution of neural networks and deep learning are intertwined, each influencing and shaping the other. This article will trace this journey, outlining the development of neural networks from their inception to their current state, and exploring how they have revolutionized the field of AI. Whether you’re a tech enthusiast keen to understand the latest trends or a novice curious about the buzzwords, this article will provide a clear and engaging overview of neural networks and their role in the exciting world of AI.
Understanding the Concept of a Neural Network
Neural networks, a subset of artificial intelligence, are computing systems inspired by the biological neural networks that constitute animal brains. These systems “learn” to perform tasks by considering examples, generally without being programmed with any task-specific rules. For instance, in image recognition, they might learn to identify images that contain cats by analyzing example images that have been manually labeled as “cat” or “no cat” and using the results to identify cats in other images.
Each ‘neuron’ in a neural network takes in inputs, which are multiplied by weights, and then passed through a nonlinear function, often called an activation function, to produce an output. The weights in the system are then adjusted based on the output, a process known as backpropagation. This is the fundamental process by which neural networks ‘learn’. The weights are adjusted to minimize the difference between the actual output and the expected output, effectively ‘training’ the network to produce the correct output for a given input.
Neural networks are organized into layers. The simplest kind is a feedforward neural network, where the information moves in only one direction—forward—from the input nodes, through the hidden nodes (if any), and to the output nodes. There are no cycles or loops in the network. More complex are recurrent neural networks (RNNs), where data can flow in any direction. RNNs are particularly effective for tasks that require ‘memory’ of prior inputs, like handwriting or speech recognition.
The power of neural networks comes from their ability to learn complex, non-linear relationships. For example, a neural network with a single layer can represent any function that separates data into two categories with a straight line or plane. However, with two layers, the network can represent an arbitrary decision boundary to arbitrary accuracy with rational activation functions and can approximate any smooth mapping to any accuracy.
Despite their power, neural networks also have limitations. They require large amounts of data to train effectively, and they are notoriously difficult to interpret. A neural network can make a prediction or classification, but it’s challenging to understand why it made that decision. This lack of interpretability is a significant drawback in many applications.
Neural networks are susceptible to adversarial attacks, where small, intentionally designed changes to input can cause the network to make a mistake. This vulnerability is a significant concern in applications where security is paramount, such as autonomous vehicles or cybersecurity. Despite these challenges, the power and flexibility of neural networks make them a valuable tool in the field of artificial intelligence.
The Biological Inspiration Behind Neural Networks
Neural networks, a cornerstone of artificial intelligence (AI), draw their inspiration from the biological systems of the brain. The human brain, composed of approximately 86 billion neurons, is a complex network of interconnected cells that transmit signals to each other, enabling us to think, learn, and react to our environment (Azevedo et al., 2009). This biological system served as the blueprint for the development of artificial neural networks (ANNs), computational models designed to mimic the brain’s ability to process and interpret information.
The fundamental unit of ANNs, the artificial neuron or node, mirrors the function of biological neurons. In the human brain, a neuron receives signals from other neurons through its dendrites, processes these signals in the cell body, and transmits the output through the axon to other neurons (Kandel et al., 2000). Similarly, an artificial neuron receives input, applies a mathematical function to process this input, and produces an output. The strength of the connections between artificial neurons, akin to the synaptic strength in biological neurons, is adjusted during the learning process, allowing the network to adapt and improve its performance over time.
The architecture of ANNs also reflects the organization of the brain. Just as neurons in the brain are arranged in interconnected layers, artificial neurons in ANNs are organized into input, hidden, and output layers. The input layer receives raw data, the hidden layers perform computations, and the output layer produces the final result. This layered structure allows ANNs to process complex, high-dimensional data, much like the human brain processes sensory information (LeCun et al., 2015).
The learning process in ANNs is inspired by the biological concept of synaptic plasticity, the ability of the connections between neurons in the brain to change in strength. The most well-known form of synaptic plasticity is Hebbian learning, often summarized as “cells that fire together, wire together” (Hebb, 1949). In ANNs, a similar principle is applied through a process called backpropagation, where the network adjusts the weights of the connections based on the error in the output (Rumelhart et al., 1986).
In conclusion, while ANNs are inspired by the biological brain, they are not an exact replica. They represent an abstraction of the brain’s complexity, designed to perform specific tasks rather than replicate the full range of human cognitive abilities. As our understanding of the brain advances, it’s likely that future models of ANNs will incorporate more biological details, leading to more powerful and versatile AI systems.
The Basic Structure of a Neural Network: Neurons and Layers
Neural networks, a subset of artificial intelligence, are designed to mimic the human brain’s ability to learn and recognize patterns. The basic structure of a neural network is composed of interconnected layers of nodes, or “neurons,” each of which performs a simple computation. The neurons are organized into three types of layers: the input layer, hidden layers, and the output layer. The input layer receives raw data, the hidden layers process the data, and the output layer produces the final result (Schmidhuber, 2015).
Each neuron in a neural network receives input from several other neurons. These inputs are each given a numerical weight, which can be adjusted to change the importance of the input. The neuron multiplies each input by its weight, then adds up all these products. This sum is then passed through a function, often a non-linear function such as a sigmoid or ReLU (Rectified Linear Unit) function, to produce the neuron’s output (Goodfellow et al., 2016).
The weights of the neurons are initially set randomly, but are adjusted as the network learns. This learning process involves feeding the network training data and adjusting the weights based on the difference between the network’s output and the desired output. This adjustment is done using a process called backpropagation, which involves calculating the gradient of the loss function with respect to the weights and then adjusting the weights in the direction that minimizes the loss (Rumelhart et al., 1986).
The layers of neurons in a neural network are interconnected, with each neuron in one layer connected to every neuron in the next layer. This fully connected structure allows the network to learn complex patterns in the data. However, it also makes the network computationally expensive to train, especially for large networks. To mitigate this, some neural networks use a convolutional structure, where each neuron is only connected to a small, local region of the previous layer. This structure is particularly effective for image recognition tasks (LeCun et al., 1998).
The number of layers in a neural network, and the number of neurons in each layer, can vary widely depending on the specific task the network is designed to perform. In general, more layers and neurons allow the network to learn more complex patterns, but also make the network more computationally expensive to train and more prone to overfitting, which is when the network learns the training data too well and performs poorly on new data (Hawkins, 2004).
In conclusion, the basic structure of a neural network consists of layers of interconnected neurons, each of which performs a simple computation. The network learns by adjusting the weights of these neurons based on the difference between its output and the desired output. The specific structure of the network, including the number of layers and neurons, can vary depending on the task the network is designed to perform.
How Neural Networks Learn: The Role of Weights and Biases
Neural networks, a subset of machine learning, are designed to mimic the human brain’s ability to learn and recognize patterns. They consist of interconnected layers of nodes, or “neurons,” which transmit information to each other much like neurons in the brain. The strength of the connections between these nodes, known as weights, and the thresholds at which they activate, known as biases, are crucial to the learning process of a neural network.
Weights in a neural network are numerical values that determine the strength and direction of the connection between two neurons. When an input is fed into the network, it is multiplied by the weight of the connection it is passing through. The higher the weight, the stronger the influence of the input on the output. During the learning process, these weights are adjusted based on the error of the network’s output compared to the desired output. This process is known as backpropagation. The goal is to minimize the error by adjusting the weights, which is typically achieved using a method called gradient descent.
Biases, on the other hand, are additional parameters in a neural network that enable it to fit the data more accurately. They allow the activation function, which determines the output of a neuron, to be shifted to the left or right, which can be crucial for successful learning. Without biases, the network would only be able to learn patterns that pass through the origin of the activation function, severely limiting its ability to generalize from the data.
The interplay between weights and biases in a neural network is a delicate balancing act. If the weights are too large, the network may overfit the data, learning the noise in the data rather than the underlying pattern. This results in poor generalization to new data. Conversely, if the weights are too small, the network may underfit the data, failing to learn the underlying pattern at all. Similarly, if the biases are set too high or too low, the network may struggle to fit the data accurately.
The process of learning in a neural network, therefore, involves iteratively adjusting the weights and biases based on the error of the network’s output. This is typically done using a set of training data, where the true output is known. The network makes a prediction based on the current weights and biases, the error is calculated, and then the weights and biases are adjusted in a direction that reduces the error. This process is repeated many times until the network’s predictions are as accurate as possible.
The Process of Training a Neural Network
The process of training a neural network, a type of machine learning model, involves several steps. The first step is the initialization of the network’s weights. These weights are typically initialized with small random numbers, which are often drawn from a Gaussian distribution (Glorot and Bengio, 2010). This randomness helps to break symmetry and prevent all neurons from learning the same features during training.
The next step in the process is the forward propagation of inputs. In this step, the input data is passed through the network, with each layer of neurons performing computations on the data and passing the results to the next layer. The computations involve multiplying the input data by the weights, adding a bias term, and then applying a non-linear activation function such as a sigmoid or ReLU function (Goodfellow et al., 2016). The output of the final layer of neurons is the network’s prediction.
The third step is the computation of the loss, which measures the discrepancy between the network’s prediction and the actual target values. Commonly used loss functions include mean squared error for regression tasks and cross-entropy for classification tasks (Bishop, 2006). The loss provides a quantitative measure of the network’s performance and serves as the objective function that the training process aims to minimize.
The fourth step is backpropagation, which is the process of computing the gradient of the loss with respect to the network’s weights. The backpropagation algorithm applies the chain rule of calculus to efficiently compute these gradients (Rumelhart et al., 1986). The gradients indicate the direction in which the weights need to be adjusted to reduce the loss.
The final step is the update of the weights, which is typically done using an optimization algorithm such as stochastic gradient descent (SGD) or one of its variants like Adam or RMSprop (Kingma and Ba, 2014). These algorithms adjust the weights in the direction of the negative gradient, thereby reducing the loss. The size of the adjustment is determined by the learning rate, a hyperparameter that controls the speed of learning.
The above steps constitute one iteration or epoch of training. The process is repeated for many epochs until the network’s performance on a validation set stops improving, at which point the training is stopped to prevent overfitting. The trained network can then be used to make predictions on new, unseen data.
Deep Learning: An Extension of Neural Networks
Deep learning, a subset of machine learning, is essentially an extension of artificial neural networks. Artificial neural networks, inspired by the biological neural networks that constitute animal brains, are computing systems designed to ‘learn’ by considering examples. They are composed of interconnected nodes, or ‘neurons’, which process information using a system of weighted connections. These weights are adjusted during the learning process to optimize the network’s performance, a process known as ‘training’ the network (Goodfellow et al., 2016).
Deep learning takes this concept and extends it by introducing multiple layers of these artificial neurons, creating what is known as a deep neural network. These layers allow the network to learn more complex patterns and relationships in the data. The first layer of the network processes raw input data, and each subsequent layer receives the output from the layer before it, refining and abstracting the information as it passes through the network. This hierarchical approach allows deep learning models to learn increasingly complex features, from simple edges and textures in early layers to complex objects and scenes in later layers (LeCun et al., 2015).
The power of deep learning lies in its ability to learn representations of data in an unsupervised manner. Traditional machine learning techniques often require manual feature extraction, where the most relevant aspects of the data are identified and extracted for use in the model. In contrast, deep learning models learn to identify these features themselves, directly from the raw data. This ability to learn useful representations from raw data is known as representation learning, and it is a key aspect of deep learning’s success in a wide range of applications (Bengio et al., 2013).
Deep learning has been particularly successful in the field of image and speech recognition. Convolutional Neural Networks (CNNs), a type of deep learning model, have achieved state-of-the-art results in image classification tasks, outperforming traditional machine learning techniques. CNNs are designed to automatically and adaptively learn spatial hierarchies of features from the raw data, making them particularly suited to image analysis (Krizhevsky et al., 2012). Similarly, Recurrent Neural Networks (RNNs), another type of deep learning model, have shown great promise in speech recognition tasks, due to their ability to process sequential data (Graves et al., 2013).
The Role of Neural Networks in the Emergence of AI
Neural networks, a subset of artificial intelligence (AI), have played a pivotal role in the emergence and advancement of AI. These networks, inspired by the biological neural networks that constitute animal brains, are designed to simulate the way in which the human brain works to solve complex problems. They are composed of interconnected nodes or ‘neurons’ that process information and learn from the data they are exposed to, thereby enabling the system to make predictions or decisions without being explicitly programmed to perform the task (Goodfellow, Bengio & Courville, 2016).
The concept of neural networks is not new. It dates back to the 1940s when Warren McCulloch and Walter Pitts proposed a computational model for neural networks based on mathematics and algorithms called threshold logic. This model laid the foundation for neural networks. However, it wasn’t until the 1980s and 1990s, when the backpropagation algorithm was introduced, that neural networks began to be widely used for AI (Schmidhuber, 2015).
The backpropagation algorithm, a method used in machine learning to train neural networks, was a game-changer. It uses a method called gradient descent to tweak the weights of neurons in a network to minimize the difference between the actual output and the desired output. This learning process is what allows a neural network to perform tasks such as image recognition, speech recognition, and natural language processing, which are central to many AI applications (Rumelhart, Hinton & Williams, 1986).
Deep learning, a more recent development in AI, is a further extension of neural networks. It involves the use of large neural networks with many layers of neurons, hence the term ‘deep’. These deep neural networks have been instrumental in achieving breakthroughs in areas such as image and speech recognition, and they are a key component of many of the AI systems we use today, from voice assistants like Siri and Alexa to recommendation systems like those used by Netflix and Amazon (LeCun, Bengio & Hinton, 2015).
However, despite their success, neural networks also have limitations. They require large amounts of data to learn effectively, and they are often described as ‘black boxes’ because it can be difficult to understand how they arrive at a particular decision. This lack of transparency can be a problem in applications where explainability is important, such as in healthcare or finance. Moreover, while neural networks are good at recognizing patterns, they are less effective at reasoning or common sense, which are areas where symbolic AI, another branch of AI, excels (Marcus, 2018).
In conclusion, neural networks have been instrumental in the emergence and advancement of AI. They have enabled breakthroughs in areas such as image and speech recognition and are a key component of many of the AI systems we use today. However, they also have limitations, and future research in AI will likely involve a combination of neural networks and other AI techniques.
Historical Evolution of Neural Networks and AI
The concept of neural networks, a cornerstone of modern artificial intelligence (AI), has its roots in the early 20th century. In 1943, Warren McCulloch, a neurophysiologist, and Walter Pitts, a logician, proposed the first mathematical model of a neural network. Their model, known as the McCulloch-Pitts (MCP) neuron, was a binary threshold neuron, meaning it could only output a signal if the input exceeded a certain threshold (McCulloch & Pitts, 1943). This model was a significant step towards the development of AI, as it provided a simplified representation of how neurons in the brain might function.
In the 1950s and 1960s, the concept of neural networks was further developed by Frank Rosenblatt, who introduced the Perceptron. The Perceptron was a more advanced model than the MCP neuron, as it could learn from its mistakes and adjust its weights accordingly (Rosenblatt, 1958). However, the Perceptron had a significant limitation: it could only solve linearly separable problems. This limitation was highlighted by Marvin Minsky and Seymour Papert in their book “Perceptrons” (1969), which led to a decline in interest and funding for neural network research.
The 1980s saw a resurgence in neural network research, largely due to the introduction of the backpropagation algorithm by Rumelhart, Hinton, and Williams (1986). Backpropagation allowed neural networks to learn from non-linearly separable data, overcoming the limitation of the Perceptron. This led to the development of multi-layer perceptrons (MLPs), which could solve more complex problems.
In the 1990s and early 2000s, support vector machines (SVMs) and other machine learning techniques became popular, overshadowing neural networks. However, neural networks made a comeback in the late 2000s and early 2010s, thanks to advances in computational power and the availability of large datasets. This period saw the rise of deep learning, a subfield of AI that focuses on training deep neural networks.
Today, neural networks and AI continue to evolve, with new architectures and techniques being developed. For example, convolutional neural networks (CNNs) have been highly successful in image recognition tasks (Krizhevsky, Sutskever, & Hinton, 2012), while recurrent neural networks (RNNs) have shown promise in natural language processing (Mikolov et al., 2010). The field of AI is now more vibrant and diverse than ever, with neural networks playing a central role.
Current Applications of Neural Networks and Deep Learning
Neural networks and deep learning, two subsets of artificial intelligence (AI), have seen a surge in application across various fields in recent years. One of the most prominent applications is in the realm of image and speech recognition. Deep learning algorithms, such as convolutional neural networks (CNNs), have been instrumental in improving the accuracy of image recognition systems. These algorithms are capable of identifying and learning patterns in large datasets, enabling them to distinguish between different images with high precision (LeCun et al., 2015). Similarly, recurrent neural networks (RNNs), another type of deep learning algorithm, have been used to enhance speech recognition systems. RNNs are particularly adept at processing sequential data, making them ideal for understanding spoken language (Graves et al., 2013).
In the healthcare sector, neural networks and deep learning have been employed to aid in disease diagnosis and treatment. For instance, deep learning algorithms have been used to analyze medical images, such as X-rays and MRIs, to detect signs of diseases like cancer (Esteva et al., 2017). Additionally, neural networks have been used to predict patient outcomes and personalize treatment plans based on individual patient data (Rajkomar et al., 2018).
The financial industry has also benefited from the application of neural networks and deep learning. These technologies have been used to predict stock market trends, assess credit risk, and detect fraudulent transactions. For example, deep learning algorithms have been used to analyze historical stock market data and predict future trends with a high degree of accuracy (Dixon et al., 2016). Similarly, neural networks have been used to assess the creditworthiness of individuals and businesses, helping financial institutions make informed lending decisions (West, 2000).
In the field of autonomous vehicles, neural networks and deep learning play a crucial role. They are used in the perception, decision-making, and control systems of these vehicles. For instance, CNNs are used to interpret visual data from cameras and sensors, enabling the vehicle to recognize objects and navigate its environment (Bojarski et al., 2016).
Finally, in the realm of natural language processing (NLP), deep learning has been used to improve machine translation, sentiment analysis, and text generation. For instance, transformer-based models like BERT (Bidirectional Encoder Representations from Transformers) have revolutionized machine translation by enabling more accurate and context-aware translations (Devlin et al., 2018).
Future Prospects: The Potential of Neural Networks in Advancing AI
Neural networks, a subset of artificial intelligence (AI), have shown immense potential in various fields, from healthcare to finance, and their future prospects are promising. These networks are designed to mimic the human brain’s structure and function, enabling machines to learn from experience, recognize patterns, and make decisions in a human-like manner (Schmidhuber, 2015). The potential of neural networks lies in their ability to learn and adapt, which is a significant advancement over traditional AI systems that rely on pre-programmed algorithms.
One of the most promising areas for the application of neural networks is in the field of healthcare. Neural networks can be trained to recognize patterns in medical images, such as X-rays or MRIs, and can potentially identify diseases or conditions that a human doctor might miss (Esteva et al., 2017). This could lead to earlier diagnosis and treatment, improving patient outcomes. Furthermore, neural networks can also be used to predict disease progression and response to treatment, aiding in personalized medicine.
In the financial sector, neural networks have the potential to revolutionize the way we manage money. They can be used to predict market trends, identify investment opportunities, and even automate trading (Dixon et al., 2017). This could lead to more efficient markets and potentially higher returns for investors. Moreover, neural networks can also be used in fraud detection, identifying suspicious patterns of behavior that might indicate fraudulent activity.
Despite these promising applications, there are still challenges to be overcome. One of the main challenges is the so-called “black box” problem: it is often difficult to understand why a neural network has made a particular decision (Castelvecchi, 2016). This lack of transparency can be problematic, especially in fields like healthcare or finance where decisions can have significant consequences. Efforts are being made to develop techniques for interpreting neural network decisions, but this remains an active area of research.
Another challenge is the need for large amounts of data to train neural networks. This can be a barrier to entry in fields where data is scarce or difficult to obtain. However, techniques for training neural networks with less data are being developed, and advances in data collection and storage may also help to alleviate this problem (Halevy et al., 2009).
References
- Rosenblatt, F. (1958). The perceptron: a probabilistic model for information storage and organization in the brain. Psychological review, 65(6), 386.
- Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I., & Fergus, R. (2013). Intriguing properties of neural networks. arXiv preprint arXiv:1312.6199.
- Hawkins, J. (2004). On Intelligence. Times Books.
- Castelvecchi, D. (2016). Can we open the black box of AI? Nature News, 538(7623), 20-23.
- Graves, A., Mohamed, A. R., & Hinton, G. (2013). Speech recognition with deep recurrent neural networks. In 2013 IEEE international conference on acoustics, speech and signal processing (pp. 6645-6649). IEEE.
- Bengio, Y., Courville, A., & Vincent, P. (2013). Representation learning: A review and new perspectives. IEEE transactions on pattern analysis and machine intelligence, 35(8), 1798-1828.
- Glorot, X. and Bengio, Y., 2010. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth international conference on artificial intelligence and statistics (pp. 249-256).
- Output References:
- Kandel, E. R., Schwartz, J. H., Jessell, T. M., Siegelbaum, S. A., & Hudspeth, A. J. (2000). Principles of neural science (Vol. 4). New York: McGraw-hill.
- Esteva, A., Kuprel, B., Novoa, R. A., Ko, J., Swetter, S. M., Blau, H. M., & Thrun, S. (2017). Dermatologist-level classification of skin cancer with deep neural networks. Nature, 542(7639), 115-118.
- LeCun, Y., Bottou, L., Bengio, Y., & Haffner, P. (1998). Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11), 2278-2324.
- Goodfellow, I., Bengio, Y., & Courville, A. (2016). Deep Learning. MIT Press.
- Nielsen, M. A. (2015). Neural Networks and Deep Learning. Determination Press.
- Schmidhuber, J. (2015). Deep learning in neural networks: An overview. Neural Networks, 61, 85-117.
- Kingma, D.P. and Ba, J., 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980.
- Halevy, A., Norvig, P., & Pereira, F. (2009). The unreasonable effectiveness of data. IEEE Intelligent Systems, 24(2), 8-12.
- Minsky, M., & Papert, S. (1969). Perceptrons: An introduction to computational geometry. MIT press.
- Rumelhart, D. E., Hinton, G. E., & Williams, R. J. (1986). Learning representations by back-propagating errors. Nature, 323(6088), 533-536.
- West, D. (2000). Neural network credit scoring models. Computers & Operations Research, 27(11-12), 1131-1152.
- Bojarski, M., Del Testa, D., Dworakowski, D., Firner, B., Flepp, B., Goyal, P., … & Zhang, X. (2016). End to end learning for self-driving cars. arXiv preprint arXiv:1604.07316.
- Azevedo, F. A. C., Carvalho, L. R. B., Grinberg, L. T., Farfel, J. M., Ferretti, R. E. L., Leite, R. E. P., … & Herculano-Houzel, S. (2009). Equal numbers of neuronal and nonneuronal cells make the human brain an isometrically scaled-up primate brain. Journal of Comparative Neurology, 513(5), 532-541.
- Devlin, J., Chang, M. W., Lee, K., & Toutanova, K. (2018). Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
- Rajkomar, A., Oren, E., Chen, K., Dai, A. M., Hajaj, N., Hardt, M., … & Sundberg, P. (2018). Scalable and accurate deep learning with electronic health records. NPJ Digital Medicine, 1(1), 1-10.
- Marcus, G. (2018). Deep learning: A critical appraisal. arXiv preprint arXiv:1801.00631.
- Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems, 25, 1097-1105.
- Dixon, M., Klabjan, D., & Bang, J. H. (2016). Classification-based financial markets prediction using deep neural networks. Algorithmic Finance, 5(3-4), 35-49.
- LeCun, Y., Bengio, Y., & Hinton, G. (2015). Deep learning. Nature, 521(7553), 436-444.
- Hebb, D. O. (1949). The organization of behavior: A neuropsychological theory. New York: Wiley.
- Mikolov, T., Karafiát, M., Burget, L., Černocký, J., & Khudanpur, S. (2010). Recurrent neural network based language model. In Eleventh annual conference of the international speech communication association.
- Bishop, C. M. (2006). Pattern Recognition and Machine Learning. Springer.
- McCulloch, W. S., & Pitts, W. (1943). A logical calculus of the ideas immanent in nervous activity. The bulletin of mathematical biophysics, 5(4), 115-133.
