Aris Xanthos
- Published in print:
- 2018
- Published Online:
- January 2019
- ISBN:
- 9780226562452
- eISBN:
- 9780226562599
- Item type:
- chapter
- Publisher:
- University of Chicago Press
- DOI:
- 10.7208/chicago/9780226562599.003.0014
- Subject:
- Linguistics, Phonetics / Phonology
This contribution describes an algorithm for the unsupervised learning of root-and-pattern morphology. The algorithm relies on a phonological heuristic to bootstrap the morphological analysis and ...
More
This contribution describes an algorithm for the unsupervised learning of root-and-pattern morphology. The algorithm relies on a phonological heuristic to bootstrap the morphological analysis and identify a preliminary set of reliable roots and patterns. The analysis is then incrementally extended based on the minimum description length principle, in line with the approach to morphological learning embodied in John Goldsmith’s Linguistica algorithm. The algorithm is implemented as a computer program named Arabica and evaluated with regard to its ability to learn the system of Arabic noun plurals. The tension between the universality of the consonant-vowel distinction and the specificity of root-and-pattern morphology turns out to be crucial for understanding the strengths and weaknesses of this approach.Less
This contribution describes an algorithm for the unsupervised learning of root-and-pattern morphology. The algorithm relies on a phonological heuristic to bootstrap the morphological analysis and identify a preliminary set of reliable roots and patterns. The analysis is then incrementally extended based on the minimum description length principle, in line with the approach to morphological learning embodied in John Goldsmith’s Linguistica algorithm. The algorithm is implemented as a computer program named Arabica and evaluated with regard to its ability to learn the system of Arabic noun plurals. The tension between the universality of the consonant-vowel distinction and the specificity of root-and-pattern morphology turns out to be crucial for understanding the strengths and weaknesses of this approach.
Ivan Herreros
- Published in print:
- 2018
- Published Online:
- June 2018
- ISBN:
- 9780199674923
- eISBN:
- 9780191842702
- Item type:
- chapter
- Publisher:
- Oxford University Press
- DOI:
- 10.1093/oso/9780199674923.003.0026
- Subject:
- Neuroscience, Sensory and Motor Systems, Development
This chapter discusses basic concepts from control theory and machine learning to facilitate a formal understanding of animal learning and motor control. It first distinguishes between feedback and ...
More
This chapter discusses basic concepts from control theory and machine learning to facilitate a formal understanding of animal learning and motor control. It first distinguishes between feedback and feed-forward control strategies, and later introduces the classification of machine learning applications into supervised, unsupervised, and reinforcement learning problems. Next, it links these concepts with their counterparts in the domain of the psychology of animal learning, highlighting the analogies between supervised learning and classical conditioning, reinforcement learning and operant conditioning, and between unsupervised and perceptual learning. Additionally, it interprets innate and acquired actions from the standpoint of feedback vs anticipatory and adaptive control. Finally, it argues how this framework of translating knowledge between formal and biological disciplines can serve us to not only structure and advance our understanding of brain function but also enrich engineering solutions at the level of robot learning and control with insights coming from biology.Less
This chapter discusses basic concepts from control theory and machine learning to facilitate a formal understanding of animal learning and motor control. It first distinguishes between feedback and feed-forward control strategies, and later introduces the classification of machine learning applications into supervised, unsupervised, and reinforcement learning problems. Next, it links these concepts with their counterparts in the domain of the psychology of animal learning, highlighting the analogies between supervised learning and classical conditioning, reinforcement learning and operant conditioning, and between unsupervised and perceptual learning. Additionally, it interprets innate and acquired actions from the standpoint of feedback vs anticipatory and adaptive control. Finally, it argues how this framework of translating knowledge between formal and biological disciplines can serve us to not only structure and advance our understanding of brain function but also enrich engineering solutions at the level of robot learning and control with insights coming from biology.
Peter Dayan and David Willshaw
- Published in print:
- 2016
- Published Online:
- January 2017
- ISBN:
- 9780198749783
- eISBN:
- 9780191831638
- Item type:
- chapter
- Publisher:
- Oxford University Press
- DOI:
- 10.1093/acprof:oso/9780198749783.003.0009
- Subject:
- Psychology, Neuropsychology
Marr’s theory of neocortex is the ghost in his trinity of seminal contributions to theoretical neuroscience, being the most spirited and least appreciated. In this chapter, we review its ...
More
Marr’s theory of neocortex is the ghost in his trinity of seminal contributions to theoretical neuroscience, being the most spirited and least appreciated. In this chapter, we review its computational and algorithmic content; others in this volume assess anatomical and physiological aspects of this theory and the related accounts of the cerebellum and the hippocampus. Marr offered an early treatment of what we would now understand as unsupervised learning in the context of how the neocortex acquires representations of sensory input. We analyze this problem, elucidating Marr’s central claims and contrasting them with intervening and current work.Less
Marr’s theory of neocortex is the ghost in his trinity of seminal contributions to theoretical neuroscience, being the most spirited and least appreciated. In this chapter, we review its computational and algorithmic content; others in this volume assess anatomical and physiological aspects of this theory and the related accounts of the cerebellum and the hippocampus. Marr offered an early treatment of what we would now understand as unsupervised learning in the context of how the neocortex acquires representations of sensory input. We analyze this problem, elucidating Marr’s central claims and contrasting them with intervening and current work.
Jackson L. Lee
- Published in print:
- 2018
- Published Online:
- January 2019
- ISBN:
- 9780226562452
- eISBN:
- 9780226562599
- Item type:
- chapter
- Publisher:
- University of Chicago Press
- DOI:
- 10.7208/chicago/9780226562599.003.0010
- Subject:
- Linguistics, Phonetics / Phonology
On what linguistic theory is about, Chomsky (1957) discussed three views: the discovery procedure, the decision procedure, and the evaluation procedure. Chomsky rejected the discovery view for being ...
More
On what linguistic theory is about, Chomsky (1957) discussed three views: the discovery procedure, the decision procedure, and the evaluation procedure. Chomsky rejected the discovery view for being “unreasonable” and argued for the evaluation view for being “practical”. Among the researchers who take the discussion seriously is John Goldsmith, with long-standing work on unsupervised learning of linguistic structure. While Goldsmith himself strongly sees his work as a response and realization of the evaluation procedure, this chapter makes the stronger claim that his research program is the higher-order goal of the discovery procedure. This chapter ends with methodological remarks on the importance of shaping linguistic research as an enterprise towards the discovery procedure.Less
On what linguistic theory is about, Chomsky (1957) discussed three views: the discovery procedure, the decision procedure, and the evaluation procedure. Chomsky rejected the discovery view for being “unreasonable” and argued for the evaluation view for being “practical”. Among the researchers who take the discussion seriously is John Goldsmith, with long-standing work on unsupervised learning of linguistic structure. While Goldsmith himself strongly sees his work as a response and realization of the evaluation procedure, this chapter makes the stronger claim that his research program is the higher-order goal of the discovery procedure. This chapter ends with methodological remarks on the importance of shaping linguistic research as an enterprise towards the discovery procedure.
Anthony Man-Cho So
- Published in print:
- 2021
- Published Online:
- April 2021
- ISBN:
- 9780198870944
- eISBN:
- 9780191913532
- Item type:
- chapter
- Publisher:
- Oxford University Press
- DOI:
- 10.1093/oso/9780198870944.003.0002
- Subject:
- Law, Intellectual Property, IT, and Media Law
Recent advances in artificial intelligence (AI) technologies have transformed our lives in profound ways. Indeed, AI has not only enabled machines to see (eg, face recognition), hear (eg, music ...
More
Recent advances in artificial intelligence (AI) technologies have transformed our lives in profound ways. Indeed, AI has not only enabled machines to see (eg, face recognition), hear (eg, music retrieval), speak (eg, speech synthesis), and read (eg, text processing), but also, so it seems, given machines the ability to think (eg, board game-playing) and create (eg, artwork generation). This chapter introduces the key technical elements of machine learning (ML), which is a rapidly growing sub-field in AI and drives many of the aforementioned applications. The goal is to elucidate the ways human efforts are involved in the development of ML solutions, so as to facilitate legal discussions on intellectual property issues.Less
Recent advances in artificial intelligence (AI) technologies have transformed our lives in profound ways. Indeed, AI has not only enabled machines to see (eg, face recognition), hear (eg, music retrieval), speak (eg, speech synthesis), and read (eg, text processing), but also, so it seems, given machines the ability to think (eg, board game-playing) and create (eg, artwork generation). This chapter introduces the key technical elements of machine learning (ML), which is a rapidly growing sub-field in AI and drives many of the aforementioned applications. The goal is to elucidate the ways human efforts are involved in the development of ML solutions, so as to facilitate legal discussions on intellectual property issues.
Kai R. Larsen and Daniel S. Becker
- Published in print:
- 2021
- Published Online:
- July 2021
- ISBN:
- 9780190941659
- eISBN:
- 9780197601495
- Item type:
- chapter
- Publisher:
- Oxford University Press
- DOI:
- 10.1093/oso/9780190941659.003.0001
- Subject:
- Business and Management, Information Technology, Innovation
Machine learning is involved in search, translation, detecting depression, likelihood of college dropout, finding lost children, and to sell all kinds of products. While barely beyond its inception, ...
More
Machine learning is involved in search, translation, detecting depression, likelihood of college dropout, finding lost children, and to sell all kinds of products. While barely beyond its inception, the current machine learning revolution will affect people and organizations no less than the Industrial Revolution’s effect on weavers and many other skilled laborers. Machine learning will automate hundreds of millions of jobs that were considered too complex for machines ever to take over even a decade ago, including driving, flying, painting, programming, and customer service, as well as many of the jobs previously reserved for humans in the fields of finance, marketing, operations, accounting, and human resources. This section explains how automated machine learning addresses exploratory data analysis, feature engineering, algorithm selection, hyperparameter tuning, and model diagnostics. The section covers the eight criteria considered essential for AutoML to have significant impact: accuracy, productivity, ease of use, understanding and learning, resource availability, process transparency, generalization, and recommended actions.Less
Machine learning is involved in search, translation, detecting depression, likelihood of college dropout, finding lost children, and to sell all kinds of products. While barely beyond its inception, the current machine learning revolution will affect people and organizations no less than the Industrial Revolution’s effect on weavers and many other skilled laborers. Machine learning will automate hundreds of millions of jobs that were considered too complex for machines ever to take over even a decade ago, including driving, flying, painting, programming, and customer service, as well as many of the jobs previously reserved for humans in the fields of finance, marketing, operations, accounting, and human resources. This section explains how automated machine learning addresses exploratory data analysis, feature engineering, algorithm selection, hyperparameter tuning, and model diagnostics. The section covers the eight criteria considered essential for AutoML to have significant impact: accuracy, productivity, ease of use, understanding and learning, resource availability, process transparency, generalization, and recommended actions.
Saul Lawrence K., Weinberger Kilian Q., Sha Fei, Ham Jihun, and Lee Daniel D.
- Published in print:
- 2006
- Published Online:
- August 2013
- ISBN:
- 9780262033589
- eISBN:
- 9780262255899
- Item type:
- chapter
- Publisher:
- The MIT Press
- DOI:
- 10.7551/mitpress/9780262033589.003.0016
- Subject:
- Computer Science, Machine Learning
This chapter provides an overview of unsupervised learning algorithms that can be viewed as spectral methods for linear and nonlinear dimensionality reduction. Spectral methods have recently emerged ...
More
This chapter provides an overview of unsupervised learning algorithms that can be viewed as spectral methods for linear and nonlinear dimensionality reduction. Spectral methods have recently emerged as a powerful tool for nonlinear dimensionality reduction and manifold learning. These methods are able to reveal low-dimensional structure in high-dimensional data from the top or bottom eigenvectors of specially constructed matrices. To analyze data that lie on a low-dimensional submanifold, the matrices are constructed from sparse weighted graphs whose vertices represent input patterns and whose edges indicate neighborhood relations. The main computations for manifold learning are based on tractable, polynomial-time optimizations, such as shortest-path problems, least-squares fits, semi-definite programming, and matrix diagonalization.Less
This chapter provides an overview of unsupervised learning algorithms that can be viewed as spectral methods for linear and nonlinear dimensionality reduction. Spectral methods have recently emerged as a powerful tool for nonlinear dimensionality reduction and manifold learning. These methods are able to reveal low-dimensional structure in high-dimensional data from the top or bottom eigenvectors of specially constructed matrices. To analyze data that lie on a low-dimensional submanifold, the matrices are constructed from sparse weighted graphs whose vertices represent input patterns and whose edges indicate neighborhood relations. The main computations for manifold learning are based on tractable, polynomial-time optimizations, such as shortest-path problems, least-squares fits, semi-definite programming, and matrix diagonalization.
Nicole Baerg
- Published in print:
- 2020
- Published Online:
- July 2020
- ISBN:
- 9780190499488
- eISBN:
- 9780190499518
- Item type:
- chapter
- Publisher:
- Oxford University Press
- DOI:
- 10.1093/oso/9780190499488.003.0006
- Subject:
- Economics and Finance, Public and Welfare
This chapter moves from studying developed countries to a sample of countries in Latin America over time. The chapter presents evidence that an increase in the information environment, in terms of ...
More
This chapter moves from studying developed countries to a sample of countries in Latin America over time. The chapter presents evidence that an increase in the information environment, in terms of its level of precision, exerts an attenuating and significant effect on the mean and standard deviation of forecasters’ inflation expectations, ultimately lowering inflation outcomes. The finding is robust to the inclusion of policy credibility, persistence in inflation, economic output, and month and country effects. When conducting instrumental variable analysis, similarly signed results hold. The main results imply that an increase in information precision helps to lower aggregate levels of inflation and that the channel that this works through is by lowering the weight of prior expectations, as predicted by the theoretical argument. Importantly, the results persist even when considering a sample of countries with relatively variable inflation outcomes and less established (and therefore less credible) economic institutions.Less
This chapter moves from studying developed countries to a sample of countries in Latin America over time. The chapter presents evidence that an increase in the information environment, in terms of its level of precision, exerts an attenuating and significant effect on the mean and standard deviation of forecasters’ inflation expectations, ultimately lowering inflation outcomes. The finding is robust to the inclusion of policy credibility, persistence in inflation, economic output, and month and country effects. When conducting instrumental variable analysis, similarly signed results hold. The main results imply that an increase in information precision helps to lower aggregate levels of inflation and that the channel that this works through is by lowering the weight of prior expectations, as predicted by the theoretical argument. Importantly, the results persist even when considering a sample of countries with relatively variable inflation outcomes and less established (and therefore less credible) economic institutions.
Nicole Baerg
- Published in print:
- 2020
- Published Online:
- July 2020
- ISBN:
- 9780190499488
- eISBN:
- 9780190499518
- Item type:
- chapter
- Publisher:
- Oxford University Press
- DOI:
- 10.1093/oso/9780190499488.003.0004
- Subject:
- Economics and Finance, Public and Welfare
Using archival, textual information from Federal Open Market Committee (FOMC) transcript data as well as FOMC policy statements, chapter 4 demonstrates that when the chair and median member have ...
More
Using archival, textual information from Federal Open Market Committee (FOMC) transcript data as well as FOMC policy statements, chapter 4 demonstrates that when the chair and median member have opposing inflation preferences, the FOMC communicates with greater precision than when the chair and median member have aligned inflation preferences. The author finds this is true, however, when computing the median members able to cast a public vote. The chapter also provides supportive evidence that when committee members are more dissimilar, the number of textual changes to the policy announcement is higher than otherwise. Theoretically, the chapter shows that a combination of members’ preferences and voting rights matter for the level of uncertainty words used in the official policy statement. Methodologically, the chapter demonstrates the innovative use of supervised and unsupervised learning techniques to construct quantitative measures from text as data, applied to central bank committees.Less
Using archival, textual information from Federal Open Market Committee (FOMC) transcript data as well as FOMC policy statements, chapter 4 demonstrates that when the chair and median member have opposing inflation preferences, the FOMC communicates with greater precision than when the chair and median member have aligned inflation preferences. The author finds this is true, however, when computing the median members able to cast a public vote. The chapter also provides supportive evidence that when committee members are more dissimilar, the number of textual changes to the policy announcement is higher than otherwise. Theoretically, the chapter shows that a combination of members’ preferences and voting rights matter for the level of uncertainty words used in the official policy statement. Methodologically, the chapter demonstrates the innovative use of supervised and unsupervised learning techniques to construct quantitative measures from text as data, applied to central bank committees.
Richard Evans
- Published in print:
- 2021
- Published Online:
- August 2021
- ISBN:
- 9780198862536
- eISBN:
- 9780191895333
- Item type:
- chapter
- Publisher:
- Oxford University Press
- DOI:
- 10.1093/oso/9780198862536.003.0011
- Subject:
- Computer Science, Artificial Intelligence
This paper describes a neuro-symbolic system for distilling interpretable logical theories out of streams of raw, unprocessed sensory experience. We combine a binary neural network, that maps raw ...
More
This paper describes a neuro-symbolic system for distilling interpretable logical theories out of streams of raw, unprocessed sensory experience. We combine a binary neural network, that maps raw sensory input to concepts, with an inductive logic programming system, that combines concepts into declarative rules. Both the inductive logic programming system and the binary neural network are encoded as logic programs, so the weights of the neural network and the declarative rules of the theory can be solved jointly as a single SAT problem. This way, we are able to jointly learn how to perceive (mapping raw sensory information to concepts) and apperceive (combining concepts into declarative rules). We apply our system, the Apperception Engine, to the Sokoban domain. Given a sequence of noisy pixel images, the system has to construct objects that persist over time, extract attributes that change over time, and induce rules explaining how the attributes change over time. We compare our system with a neural network baseline, and show that the baseline is significantly outperformed by the Apperception Engine.Less
This paper describes a neuro-symbolic system for distilling interpretable logical theories out of streams of raw, unprocessed sensory experience. We combine a binary neural network, that maps raw sensory input to concepts, with an inductive logic programming system, that combines concepts into declarative rules. Both the inductive logic programming system and the binary neural network are encoded as logic programs, so the weights of the neural network and the declarative rules of the theory can be solved jointly as a single SAT problem. This way, we are able to jointly learn how to perceive (mapping raw sensory information to concepts) and apperceive (combining concepts into declarative rules). We apply our system, the Apperception Engine, to the Sokoban domain. Given a sequence of noisy pixel images, the system has to construct objects that persist over time, extract attributes that change over time, and induce rules explaining how the attributes change over time. We compare our system with a neural network baseline, and show that the baseline is significantly outperformed by the Apperception Engine.